 Research
 Open access
 Published:
Albatross analytics a handson into practice: statistical and data science application
Journal of Big Data volume 9, Article number: 70 (2022)
Abstract
Albatross Analytics is a statistical and data science data processing platform that researchers can use in disciplines of various fields. Albatross Analytics makes it easy to implement fundamental analysis for various regressions with random model effects, including Hierarchical Generalized Linear Models (HGLMs), Double Hierarchical Generalized Linear Models (DHGLMs), Multivariate Double Hierarchical Generalized Linear Models (MDHGLMs), Survival Analysis, Frailty Models, Support Vector Machines (SVMs), and Hierarchical Likelihood Structural Equation Models (HSEMs). We provide 94 types of dataset examples.
Introduction
The application of statistical data processing has grown during the last decade, starting from traditional methods to advanced methods, including machine learning and extensive data analysis. The objective of statistical inference is to draw conclusions about a study population based on a sample of observations. Recently, subjective specific beliefs have been developed by introducing random effects in various components of models [1]. Different study problems involve specific sampling techniques and a statistical model to describe the analyzed situation.
Albatross Analytics is a statistical and computational data analysis program belonging to the opensource software class built after the R program package with the S programming language. Albatross Analytics is currently under a project by HGLM’s worldwide group. In particular, it provides a new unified stateoftheart statistical package from basic analysis to advance analysis including various randomeffect models (HGLMs, DHGLMs, MDHGLMs, and frailty models) whose implementations are generally difficult.
Meanwhile, the basis of Albatross analytics in R software is clear R software was first worked on by Robert Gentleman and Ross Ihaka of the University of Auckland's Statistics Department in 1995 [2, 3]. Most of the functionality and capabilities of Albatross Analytics can be obtained through Add—packages/libraries.
A library is a collection of commands or functions that can perform specific analyses. For instance, this implements double hierarchical generalized linear models in which the mean, dispersion parameters for the variance of random effects, and residual variance (overdispersion) can be further modeled as a randomeffect model may use DHGLM [4], MDHGLM by Lee [5,6,7]. This package allows various models for multivariate response variables where each response is assumed to follow double hierarchical generalized linear models. See also further HGLM applications for machine learning [4], schizophrenic behavior data [8], variable selection methods [9], nonGaussian factor [10], factor analysis for ordinal data [11], survival analysis [12], longitudinal outcomes and timetoevent data [13], and recent advanced topics [14,15,16,17].
The FRAILTYHL package fits semiparametric frailty and competing risk models using the hlikelihood. This package allows lognormal or gamma frailties for randomeffect distribution, and it fits shared or multilevel frailty models for correlated survival data. Functions are provided to format and summarize the FRAILTYHL results [18]. The estimates of fixed effects and frailty parameters and their standard errors are calculated. We illustrate the use of our package with two wellknown data sets and compare our results with various alternative Rprocedures. Refers to the application of semicompeting risks data [19], and clustered survival data [20, 21]. This paper addresses and explains what Albatross Analytics is and include how to use it in statistical and data science application. The advantage of Albatross Analytics is the user can analyze and interpret the data easily. Meanwhile, Fig. 1 shows the feature of Albatross Analytics, including fundamental analysis, random effect, regression, survival analysis, and multiple response analysis. This paper aims to express the application of Albatross Analytics software to handle statistical analysis in broad areas. Long story short, we provide illustrative examples. A handson various applications including HGLM, DHGLM, MDHGLM, Survival Analysis, Frailty Models, Support Vector Machine, and Structural Equation Models.
Illustrative examples
Data management
In today's world, data is the driving factor behind all establishments. As institutions keep collecting so much data, there is a need to handle the quality of the data becoming more notable by the day. Data Quality Management is the set of measures applied by a technical team or a database management system to enable good new knowledge [22,23,24]. The above collection of techniques is decided to carry out during the data management pathway, from data capture to execution, dissemination, and interpretation [24,25,26]. In line with this, the data management is the process of processing, managing, and maintaining data quality [27, 28]. Effective data management can increase the efficiency of research work [26, 29]. Figure 2a describes the main features available in Albatross Analytics. In the import data section, users can maximize this feature to upload data to be processed where the possible files are in excel and txt formats, respectively. For instance, Fig. 2b explains how to make a new variable feature, merge the dataset, and add new variables.
Each expression or variable has a data type such as numeric, integer, complex, logical, and character. The data types in Albatross analytics are expressed in class. A class is a combination of data types and the operations performed against the dataset type. The Albatross analytics look at the data as objects having attributes or properties. Data properties are defined by data type.
Basic analysis and GLMs
Descriptive statistics have been used to identify the specific characteristics of the data in the interpretation. We provide simple details of the findings and the procedures followed in Fig. 3. Alongside primary frequency distribution, we form the basis of almost all quantitative analyses of the results. Descriptive statistics shall be used to present practical explanations understandably. Descriptive statistics allow one to interpret enormous amounts of data in a structured way.
The ttest can be used to compare the means of two groups of data with the type of interval scale variable. Sometimes We will come across a study that aims to compare the mean of a sample with the mean of the entire population. Research models like this are rare, but the researcher can still provide valuable assumptions. We can do two kinds of tests, including ztest and a ttest. The condition we need to pay attention to is the population's standard deviation. If we know the standard deviation, we get it using the ztest. This will be found very rarely or never. Therefore, the most frequently used test is the ttest because we do not need to know the standard deviation of the population we study.
Furthermore, the use of the ttest on two samples is divided into two types based on the characteristics of the two samples. The first is the ttest on two independent samples. This means that the two samples to be studied came from two different groups and were given further treatment.
During the research, the use of analysis of variance is fundamental. One of the assumptions that must be met is that the population variances are the same, so we need to test the hypothesis. The purpose of the analysis of variance (ANOVA) is to determine the similarity of several population means. Oneway ANOVA may be used if only one factor is involved. Two types of tests can be used in ANOVA testing, including formal and visual tests.
Meanwhile, the statistical test can be conducted by model checking plot. If the plot does not form a specific pattern, it is said that the homogeneity of the variance is fulfilled. We know the characteristics of each variable using descriptive analysis. In addition, we may see the relationship between variables, either normal or nonnormal data [30]. With this correlation test, we want to know the similarity of the trends of the two variables. When the value of variable increases, it will also be accompanied by an increase or decrease in the value of other variables [31].
One main factor determines the test method used, namely the distribution of the data to be tested. We can use the parametric correlation test if the data distribution is normal, including Pearson's correlation coefficient. Besides, if the data distribution is not normal, we can use Kendall's rank correlation and Spearman’s rank correlation, which are nonparametric correlation tests.
Regression analysis tests the causal relationship between variables—one variable as the independent variable and one other variable as the dependent variable. Numerous regression approaches, including Poisson regression, were used during the 1970s. Linear regression and logistic regression require a unique estimation algorithm by maximizing the likelihood. Figure 4 explains that Albatross Analytics provides features for using the Linear model, GLM Logit Model, GLM Probit Model, Loglinear Model, and joint GLM.
GLM describes a family of models where the response comes from the exponential family of distributions. The method used to ttest or Ftest and inferences of these models is maximum likelihood (ML). In the GLM family of models, an IWLS algorithm can compute the ML estimates and their standard errors. Hence, the computational machinery developed for leastsquares estimation for linear models can fit GLMs, but the statistical method is based on ML.
Handson and application albatross analytics
Hierarchical generalized linear models (HGLMs)
Albatross Analytics’ distinct advantage is its unified analysis of random effect models. Various random effect models can be represented as HGLMs and estimated by hlikelihood procedures [32]. HGLMs are defined as follows:

(1)
Conditional on random effects \(u\), the responses \(y\) follows a GLM family, satisfying
$$E\left( {yu} \right) = \mu \,{\text{and}}\,var\left( {yu} \right) = \phi V\left( \mu \right),$$for which the kernel of the loglikelihood is given by
$$\sum {{{\left\{ {{\text{y}}\theta  {\text{b}}\left( \theta \right)} \right\}} \mathord{\left/ {\vphantom {{\left\{ {{\text{y}}\theta  {\text{b}}\left( \theta \right)} \right\}} {\phi ,}}} \right. \kern\nulldelimiterspace} {\phi ,}}}$$where \(\theta =\theta (\mu )\) is the canonical parameter. The linear predictor takes the form in Eq. (1):
$$\eta =g\left(\mu \right)=X\beta +Zv,$$(1)where \(v=v(u)\), for some monotone function \(v(\cdot )\) and the link function \(g\left(\mu \right)\).

(2)
The random component \(u\) follows a (conjugate) distribution to a GLM family of distributions with parameter \(\lambda\).
To infer the HGLM, Lee and Nelder [32] proposed using the hlikelihood. The h (log) likelihood is defined as Eq. 2:
The GLM attributes of an HGLM are summarized in Fig. 4.
In Bissell’s fabric study, the response variable \(y\) is the number of faults in a bolt of the fabric of length \(l\). Table 1 represents the results of the fabric study. Figure 6 illustrates the negative binomial model fitted via Poissongamma HGLM with saturated random effects for the complete response. In addition, the model checking plot is presented in Fig. 5.
Double hierarchical generalized linear models (DHGLMs)
HGLM can be extended by allowing additional random effects in their various components. Lee and Nelder [32] introduced a class of double HGLMs (DHGLMs) in which random effects can be specified in both the mean and the residual variances. Heteroscedasticity between clusters can be modeled by introducing random effects in the dispersion model as heterogeneity between clusters in the mean model. With DHGLMs, it is possible to have robust inference against outliers by allowing heavytailed distribution. Many models can be unified and extended further by the use of DHGLMs. These also include models in the finance area such as autoregressive conditional heteroscedasticity (ARCH) models, generalized ARCH (GARCH), and stochastic volatility (SV) models. Models can be further extended by introducing random effects in the variance terms. Suppose that conditional on the pair of random effects \((a, u)\), the response \(y\) satisfies.
The critical extension is to introduce random effects into the component \(\phi\):

(1)
Given \(u\), the linear predictor for \(\mu\) takes the HGLM form in Eq. 1 where \(g(\cdot )\) is the link function, \(X\) and \(Z\) are model matrices, \(v={g}_{M}\left(u\right)\) for some monotone function, \({g}_{M}\left(u\right)\) are the random effects, and \(\beta\) are the fixed effects. Moreover, dispersion parameters \(\lambda\) for \(u\) have the GLM form in Eq. 3
$${\xi }_{M}={h}_{M}\left(\lambda \right)={G}_{M}{\gamma }_{M},$$(3)where \({h}_{M}()\) is the link function, \({G}_{M}\) is the model matrix and \({\gamma }_{M}\) is fixed effects.

(2)
Given \(a\), the linear predictor for \(\phi\) takes the HGLM form as described in Eq. 4
$$\xi =h\left(\phi \right)=G\gamma +Fb,$$(4)
where \(h()\) is the link function, \(G\) and \(F\) are model matrices, \(b={g}_{D}\left(a\right)\) for some monotone function, \({g}_{D}\left(a\right)\) are the random effects, and \(\gamma\) are the fixed effects. Moreover, dispersion parameters \(\alpha\) for \(a\) have the GLM form, as shown in Eq. 5.
where \({h}_{D}(\cdot )\) is the link function, \({G}_{D}\) is the model matrix and \({\gamma }_{D}\) is fixed effects. Here, the labels \(M\) and \(D\) stand for mean and dispersion, respectively. The GLM attributes of a DHGLM are summarized in Fig. 4.
However, We illustrate an example of how to fit the DHGLM. Hudak [33] presented crack growth data, listed in Lu [34]. Each of 21 metallic specimens was subjected to 120,000 loading cycles, with the crack lengths recorded every 10,000 cycles. Let \({l}_{ij}\) be the crack length of the \(i\)th specimen at the \(j\)th observation and \({y}_{ij}={l}_{ij}{l}_{ij1}\) be the corresponding increment of crack length (response variable) measured in inches, which always has a positive value. A detailed description of the model can be found in Table 2, and Fig. 5a and b represent the mean and the dispersion, respectively [5]. Compared to an HGLM, DHGLM gives model checking plots for mean and dispersion, respectively.
Multivariate double hierarchical generalized linear models (MDHGLM’s)
Using hlikelihood, multivariate models are directly extended by assuming correlations among random effects in DHGLMs for different responses. The use of hlikelihood indicates that interlinked GLM fitting methods for HGLMs can be easily extended to fit multivariate HGLMs (MDHGLMs). Moreover, the resulting algorithm is numerically efficient and gives statistically valid inferences. In this paper, we present the example for MDHGLM. For more details, see [35] Meanwhile, Price et al. [36] presented data from a study on the developmental toxicity of ethylene glycol (EG) in mice. Table 3 summarizes the data on malformation (binary response) and fetal weight (continuous response) and shows clear doserelated trends concerning both responses.
To fit the EG data, the following bivariate HGLM is considered:

(1)
\({y}_{1ij}{w}_{i}\sim N\left({\mu }_{ij}, \phi \right), {\mu }_{ij}={x}_{1ij}{\beta }_{1}+{w}_{i}\),

(2)
\({y}_{2ij}{u}_{i}\sim Ber\left({p}_{ij}\right), logit({p}_{ij})={x}_{2ij}{\beta }_{2}+{u}_{i}\), and

(3)
\({({w}_{i}, {u}_{i})}^{T} \sim BVN\left(\mathbf {0,{\varvec{\Sigma}}}\right), cor({w}_{i},{u}_{i})=\rho\).
Figure 6 shows the path diagram of the model for the EG data. The malformation model information is given in Table 4, with cAIC for the evaluation models. In line with this, we get the result for the weight model in Table 5 and correlation in Table 6.
Survival analysis
Albatross Analytics also provides features for survival analysis, which represent in Fig. 7 by including incomplete data caused by censoring in survival time (timetoevent) data including Kaplan–Meier Estimator, Cox Model, Frailty Model [7], and Competing Risk Model [19, 37]. More instances of the Kaplan Meier curve describe the relationship between the estimated survival function at time t and the survival time. The vertical axis represents the estimated survival function, and the horizontal axis represents the survival time.
Cox proportional hazards (PH) regression is used to describe the relationship between the hazard function of survival time and independent variables which are considered to affect survival time. Cox regression is a common regression used in survival analysis because it does not assume a particular statistical distribution (e.g., baseline hazard) of the survival time.
Cox’s PH model is widely used to analyze survival data. This method is helpful with its semiparametric existence, whereby baseline hazards are nonparametric, and treatment effects are estimated parametrically. A partial likelihood has usually been used to accommodate such a semiparametric form. However, it can also be fitted with Poisson GLM methods. Moreover, they are sluggishly led to many nuisance parameters induced by nonparametric measurement hazards. Meanwhile, using the hlikelihood theory, we can prove that Poisson HGLM methodologies could be used for such kinds of modeling techniques. That being said, this method is again sluggish since the number of nuisance parameters in nonparametric baseline hazards grows with the number of events.
Example 1 using incomplete data caused by censoring in survival data
In Fig. 7, we study the analysis of incomplete data caused by censoring survival data. Cox’s PH model is widely used to analyze survival data. Frailty models with a nonparametric baseline hazard extend the PH model by allowing random effects in hazards and have been widely adopted for the analysis of correlated or clustered survival data using hlikelihood theory; we can show that Poisson HGLM algorithms can be used to fit the frailty models [12, 38,39,40,41,42,43].
Data consist of rightcensored observation from q subjects, with \({n}_{i}\) observations each (\(i=1,\dots ,q)\), \(n={\Sigma }_{i}{n}_{i}\) as the total sample size, \({T}_{ij}\) as survival time for the \(j\)th observation of the ith subject (\(j=1,\dots ,{n}_{i})\), \({C}_{ij}\) as the corresponding censoring time, \({y}_{ij}=\mathrm{min}\left\{{T}_{ij},{C}_{ij}\right\}, {\delta }_{ij}=I({T}_{ij}\le {C}_{ij})\), and \({u}_{i}\) as observed frailty for the ith subject. The conditional hazard function of \({T}_{ij}\) given \({u}_{i}\) is of the form in Eq. 6
Here \({\lambda }_{0}\left(\cdot \right)\) is an unspecified baseline hazard function and \(\beta ={\left({\beta }_{1},\dots ,{\beta }_{p}\right)}^{T}\) is a vector of regression parameters for the fixed covariates \({x}_{ij}\). Here, the term \({x}_{ij}^{T}\beta\) does not include an intercept term because of identifiability. Then, we assume that the frailties \({u}_{i}\) are i.i.d random variables with a frailty parameter \(\alpha\). We often assume gamma or lognormal distribution for \({u}_{i}\); that is, it is gamma frailty with \(E\left({u}_{i}\right)=1\) and \(\mathrm{var}\left({u}_{i}\right)=\alpha\) and lognormal frailty with \({v}_{i}=\mathrm{log}{u}_{i}\sim N(0,\alpha )\). Meanwhile, the multicomponent frailty models can be expressed in Eq. 7, with the linear predictor
\(X\) is \(n\times p\) model matrix for \(\beta\), \(\mathrm{and }{Z}^{r}\) is \(n\times {q}_{r}\) model matrices corresponding to the frailties \({v}^{r}\). At the same time, \({v}^{(r)}\) and \({v}^{(i)}\) are independent for \(r\ne I.\) Also, \({Z}^{r}\) has indicator values such that \({Z}_{st}^{(r)}=1\) if observation \(s\) is a member of the subject \(t\) in the \(r\)th frailty component, and 0 otherwise.
To the illustration, below we present two examples. Example 1 considers the dataset of the recurrence of infections in kidney patients using a portable dialysis machine. The data consist of the first and second recurrences of kidney infection in 38 patients. The catheter is later removed if the condition occurs and can be removed for other reasons, which we regard as censoring (about 24%).
In Example 1, the variables consist of 38 patients (id), time until infection since the catheter insertion (time), and a censoring indicator (1, infection; 0, censoring) for status, age of the patient (age), sex (sex) of the patient (1, male; 2, female), disease types (disease) following GN, AN, PKD, other, and estimated frailty (frail). The survival times (1st and 2nd infection times) for the same patient are likely to be correlated because of shared frailty describing the common patient’s effect. We thus fit lognormal frailty models with two covariates, sex, and age. Here, we consider the patient as frailty. Figure 8 presents the Kaplan–Meier plot for the estimated survival probability of the sex (sex1, male; sex2, female). This shows that the female group has overall higher survival (i.e., less infectious) probabilities than ones in the male group. Table 7 summarizes the estimated results of the lognormal frailty model. We show the estimated frailty in Fig. 9. For further discussions in survival analysis, see [18].
Example 2 placebocontrolled rIFNg in the treatment of CGD
Example 2, in the following case examples, consists of a placebocontrolled rIFNg in the treatment of CGD [44, 45]. One hundred twentyeight patients from 13 centres were tracked for around 1 year. The survival times are the recurrent infection times of each patient. Censoring occurred at the last observation for all patients, except one, who experienced a severe infection on the date he left the study. About 63% of the data were censored. The recurrent infection times for a given patient are likely to be correlated. Also, each patient belongs to one of the 13 centres. The correlation may be attributed to the patient effect and centre effect. Meanwhile, the recurrent infection times of each patient or censoring time (tstart–tstop), 128 patients (id), 13 centers (center), rIFNg or placebo (treat), censoring indicator (1, infection observed; 0, censored) for status, data of randomization (random) information about patients at study entry (sex, age, height, weight), the pattern of inheritance (inherit), use of steroids at study entry 1(yes), 0(no) (steroids), use of propylac antibiotics at study entry. 1(yes), 0(no) (propylac), categorization of the centers into four groups (hos.cat), and observation number withinsubject (enum). We fit multilevel lognormal frailty with two frailties and a single covariate, treatment. Here, the two frailties are random center and patient terms, with their structures given in Eq. 8.
Here \({v}^{1}\) is center frailty, and \({v}^{2}\) is patient frailty. For testing the need for a random component i.e.,\({(\alpha }_{1}=0 \, \mathrm {or}\,{\alpha }_{2}=0)\) we use the deviance \(\,2{p}_{\beta ,\mathrm{v}}\left({h}_{p}\right)\) and fit the following four models.
M1 Cox’s model without frailty \({(\alpha }_{1}=0\, \mathrm{or}\,{\alpha }_{2}=0):\,2{p}_{\beta ,\mathrm{v}}\left({h}_{p}\right)=707.48\)
M2 model without patient effect \({(\alpha }_{1}>0\, \mathrm{or}\,{\alpha }_{2}=0):\,2{p}_{\beta ,\mathrm{v}}\left({h}_{p}\right)=703.66\)
M3 model without center effect \({(\alpha }_{1}=0\, \mathrm{or}\,{\alpha }_{2}>0):\,2{p}_{\beta ,\mathrm{v}}\left({h}_{p}\right)=692.99\)
M4 multilevel model \({(\alpha }_{1}>0 \mathrm{or} {\alpha }_{2}>0):2{p}_{\beta ,\mathrm{v}}\left({h}_{p}\right)=692.95\).
Table 8 represents the model description. The deviance difference (692.99 − 692.95 = 0.04) between M3 and M4 \((0.04<2.71={\chi }_{0.10}^{2} \left(1)\right)\) indicates the absence of the random center effects, and the deviance difference between M2 and M4 (10.71) shows the necessity of random patient effects. In addition, the deviance difference between M1 and M3 (14.49) presents the random patient effect with or without random center effects. All of the three criteria (cAIC, mAIC and rAIC) also choose M3 among the M1–M4. Figure 10 presents the estimated frailty effects of this study. The explanations of model evaluation toward these three criteria can be seen in the Appendix.
Support vector machine using H likelihood
Support Vector Machine (SVM) is a supervised learning method for classification and regression using nonlinear boundaries by feature space [4, 46,47,48,49]. We present a Support Vector Machine (SVM) based on the HGLM method [4]. The match between the observed response and the model output is optimized. The output model is a feature or prognostic function also referred to as a utility function and more specifically in medical research it is called the prognostic index or health function, defined in Eq. 9:
Here \(u:{\mathbb{R}}^{d}\to {\mathbb{R}}\), \(w\) is a vector of unknown \(d\) parameters and \(\varphi \left(x\right)\) is the transformation of the covariates x. In nonlinear SVM, the transformation function used is "Kernel Trick", see: [50,51,52] Kernel Trick calculates the scalar product in the form of a kernel function. The SVM model is implied with a constraint function that will get the right margin. The constraint function of the SVM model is shown in Eq. 10. If there is an error in ranking it is given by the slack variable \({\xi }_{ij}\ge 0\). The formulation of the SVM model is described in Eq. 10: cantered depression, and the latent person
with a regularization parameter \(\upgamma \ge 0\). \({v}_{ij}\) is an indicator function of whether or not two subjects with observations \(i\) and \(j\) are comparable; it is 1 if \(i\) and \(j\) are comparable and 0 otherwise. In this paper, we use the dataset of the anatomy of an Abdominal Aortic Aneurysm (AAA), Aortic Anatomy on Endovascular Aneurysm Repair (EVAR), see [53]. The variables are described as follows: Y = Sex, X_{1} = Age, X_{2} = Aortic type Fusiform (1), Saccular (2), X_{3} = Proximal neck length, X_{4} = Proximal neck diameter, X_{5} = Proximal neck angle, and X_{6} = Max. Aneurysmal sac. We set the response variable towards simulation by following the Bernoulli distribution with 500 observations. In each scenario, the process of generating data is repeated 100 times. The parameter values used are: \(\gamma =0.7, Cost=8\). Verbose takes advantage of a perprocess runtime setting. Meanwhile, the SVM parameter setting is as follows:

First simulation:
Cluster method: “kmeans”, cost = 8, lambda = 1, centers = 2, verbose = 0.

Second simulation:
Cluster method: “kernkmeans”, cost = 8, lambda = 1, centers = 2, verbose = 0.

Third simulation:
Cluster.method = "kernkmeans", cost = 8, lambda = 1, centers = 3, verbose = 0.

Fourth simulation:
Cluster.method = "kernkmeans", cost = 8, lambda = 1, centers = 4, verbose = 0.
There are two types of model evaluation criteria: the classification stage and the HGLM analysis stage. Evaluation of the model’s goodness at the classification stage uses AUC and is determined using the values contained in the confusion matrix, with
This simulation shows that HGLM performs better with high sensitivity because some of the data used is a binary case that SVM cannot handle. For more information on step construction using hierarchical likelihood towards SVM. Table 9 represents that the use of Ensemble SVM reduces the accuracy and other measures. When the mixture patterns exist in the predictor, Ensemble SVM improves SVM performance in two scenarios. Ensemble SVM performed almost as well as logistic regression, except for sensitivity. There is a decrease in performance in the Ensemble SVM model in the multicollinearity condition and linear combination between the predictor variables. Meanwhile, HGLM still has a good performance, which is represented in Fig. 11a and b, respectively.
Using Hlikelihood to structural equation model (HSEMs)
The is widely used in multidisciplinary fields [41]. To account for the information, [42, 43] performs the style of frequentist model averaging in structural equation modeling, nonlinear structural equation modeling towards ordinal data [44] and partial least square [45, 46] and robust nonlinear with the interaction between exogenous and endogenous latent variables [47]. With an example we present a SEM method based on hlikelihood, called “hsem” [52].
In application, [48] uses twolevel dynamic SEM on longitudinal data at Mplus. In this paper, we explicitly discuss how to use hLikelihood in SEM. This data set consists of 50 repetitions on regular time scales for 100 individuals. For the response variable, the urge to smoke is on a standardized scale so that 0 corresponds to the average where the standard deviation is 1. Smokers can feel drastic mood changes. Starting from feeling happy then turning into sadness, this can show the characteristics of a person who is depressed. For those addicted, smoking can give a calm mind for a moment. The second model will answer the question; latent person predicts smoke, mean cantered depression, and the latent personmean centered lag1 urge to smoke. The model Eq. 11 is given as follows:
Figure 12 represents the path diagram by using hsem. This same standard progression path across all respondents was defined through the fixedeffect model. In contrast, the personspecific random effects are used to catch the variance of each participant from the expected path. Meanwhile, the path diagram represents withinlevel and betweenlevel models. As more instance, we provide the R package hsem [54].
Short review albatross analytics
This paper explains how Albatross software can be used for alternative multidisciplinary data processing. We offer model estimation, model checking plots, and visualization features to interpret information. Through data and R code, instances would further reveal the benefit of the HGLM model for particular statistical cases. The hlikelihood approach is distinct from both classical frequentist and Bayesian frameworks, while this encompasses inference of both fixed and random unknowns. The main benefit over classical frequentist approaches is that it would be possible to infer unobservable quantities, such as random effects, and therefore, observations could be rendered. Whenever a statistical model has been selected for the research, the likelihood contributes to the direction of inferential statistics.
Throughout direct ties with the establishment of hlikelihood, the nomenclature has already been used in which a wide variety of likelihoods have now been established. Most are through theoretical computation of GLM and GLMM, e.g., quasilikelihood and extended quasilikelihood. Many others are used to show the linkage of conventional frequentist estimation and Bayesian inference by the following other terms such as joint likelihood, extended likelihood, and adjusted profile likelihood. We demonstrate whether hlikelihood was an essential likelihood that marginal and REML probabilities and statistical probabilities are extracted. The extended probability theory underlies the hlikelihood system and demonstrates how it holds from classical and Bayesian probability.
Generalizations on random effects are of great application in simulations. For example, a typical example is that there are frequent observations of hospital admissions by patients and that the life of these patients can be expected. This might include a survival experiment with unexpected results for patients, and the variance of the estimates indicates the variability of the random effect.
During the first few examples, we demonstrate experiments using normal, lognormal, gamma, Poisson, and binomial HGLMs. Binary models are used to compare with application areas, while the dhglm package is fast and yields consistent results. Descriptions using HGLMs, including organized dispersion, are given below. We also line up models including correlated random effects and structural equation models.
The likelihood implies that probability models will offer an effective way to interpret the data if the model is accurate. It is also necessary to validate the model to verify the interpretation of the results. That being said, it could be hard to ascertain all the model assumptions. During the simulation using hlikelihood, SEM’s normal assumption in binary GLMMs can give serious biases if the normal assumption on random effects is incorrect.
Conclusion and future research
The likelihood inferences for specific models may be susceptible to data leakage outliers. If the data size is limited, we can review the data carefully to detect outliers, but it can be difficult for largescale data to identify outliers or degraded data. A commonly cited drawback of the probability approach is that it is not resilient to model distribution predictions or the existence of outliers or data degradation. It is advantageous to build models that are likely to have stable inferences against such violations. That’s also feasible by believing that the model encompasses a wide variety of distributions. We are leaving future studies to combine hlikelihood in the deep learning [39, 40, 55,56,57,58,59], and using this framework towards spatial and remote sensing [60,61,62,63,64], hybrid forecasting [65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80], and more advanced disease detection cases using image detection [81,82,83,84,85,86,87,88,89,90].
Data availability
The analysis codes and datasets used in this paper are available from the corresponding author upon reasonable request. Also, the reader can reach the Albatross Analytics website http://cheoling.snu.ac.kr:3838/DHGLM/ to perform graphical and statistical analysis.
Abbreviations
 AAA:

Abdominal aortic aneurysm
 ANOVA:

Analysis of variance
 ARCH:

Autoregressive conditional heteroskedasticity
 AUC:

Area under curve
 cAIC:

Conditional akaike information criterion
 CGD:

Chronic granulomatous disease
 DHGLM:

Double hierarchical generalized linear models
 EG:

Ethylene glycol
 EVAR:

Aortic anatomy on endovascular aneurysm repair
 frailtyHL:

Frailty models via hierarchical likelihood
 GLM:

Generalized linear model
 HGLM:

Hierarchical generalized linear models
 IWLS:

Iterated weighted least squares
 mAIC:

Marginal Akaike information criterion
 mdhglm:

Multivariate double hierarchical generalized linear models
 ML:

Maximum likelihood
 rAIC:

Restricted Akaike information criterion
 REML:

Restricted maximum likelihood
 rIFNg:

Randomized trial of gamma interferon
 SEM:

Structural equation model
 SV:

Stochastic volatility
 SVM:

Support vector machine
References
Lee Y, Rönnegård L, Noh M. Data analysis using hierarchical generalized linear models with R. 1st ed. Florida: Routledge; 2017.
R Development Core Team. R: a language and environment for statistical computing. Vienna: R Foundation for Statistical Computing; 2011.
R Core Team. R software. Vienna: R Foundation for Statistical Computing; 2008. p. 409. https://doi.org/10.1007/9783540746867.
Caraka RE, Lee Y, Chen RC, Toharudin T. Using hierarchical likelihood towards support vector machine: theory and its application. IEEE Access. 2020;8:194795–807.
Lee Y, Rnnegrd L, Noh M. Double HGLMs using the dhglm package. In: Noh M, editor. Data analysis using hierarchical generalized linear models with R. London: Chapman and Hall; 2017.
Felleki M, Lee D, Lee Y, Gilmour AR, Rönnegård L. Estimation of breeding values for mean and dispersion, their variance and correlation using double hierarchical generalized linear models. Genet Res. 2012;94:307–17.
Ha ID, Noh M, Lee Y. FrailtyHL: a package for fitting frailty models with hlikelihood. R J. 2012;4:28–37.
Lee Y, Noh M. Modelling random effect variance with double hierarchical generalized linear models. Stat Model. 2012;12:487–502.
Collignon O, Han J, An H, Oh S, Lee Y. Comparison of the modified unbounded penalty and the LASSO to select predictive genes of response to chemotherapy in breast cancer. PLoS ONE. 2018;13:15.
Noh M, Lee Y, Oud JHL, Toharudin T. Hierarchical likelihood approach to nonGaussian factor analysis. J Stat Comput Simul. 2019;89:1555–73.
Jin S, Noh M, Lee Y. Hlikelihood approach to factor analysis for ordinal data. Struct Equ Model. 2018;25:530–40.
Ha ID, Lee Y. A review of hlikelihood for survival analysis. Jpn J Stat Data Sci. 2021. https://doi.org/10.1007/s4208102100125z.
Ha ID, Noh M, Lee Y. Hlikelihood approach for joint modeling of longitudinal outcomes and timetoevent data. Biom J. 2017;59:1122–43.
Lee D, Lee Y. Extended likelihood approach to multiple testing with directional error control under a hidden Markov random field model. J Multivar Anal. 2016;151:1–13.
Lee W, Ha ID, Noh M, Lee D, Lee Y. A review on recent advances and applications of hlikelihood method. J Korean Stat Soc. 2021. https://doi.org/10.1007/s42952021001308.
Jin S, Lee Y. A review of hlikelihood and hierarchical generalized linear model. WIREs Comp Stat. 2020. https://doi.org/10.1002/wics.1527.
Caraka RE, Noh M, Chen RC, Lee Y, Gio PU, Pardamean B. Connecting climate and communicable disease to penta helix using hierarchical likelihood structural equation modelling. Symmetry. 2021;13:1–21.
Ha ID, Jeong JH, Lee Y. Statistical modelling of survival data with random effects. Berlin: Springer; 2017.
Ha ID, Xiang L, Peng M, Jeong JH, Lee Y. Frailty modelling approaches for semicompeting risks data. Lifetime Data Anal. 2020;26:109–33.
Huang R, Xiang L, Ha ID. Frailty proportional mean residual life regression for clustered survival data: a hierarchical quasilikelihood method. Stat Med. 2019;38:4854–70.
Ha ID, Kim JM, Emura T. Profile likelihood approaches for semiparametric copula and frailty models for clustered survival data. J Appl Stat. 2019;46:2553–71.
Taleb I, Serhani MA, Bouhaddioui C, Dssouli R. Big data quality framework: a holistic approach to continuous quality management. J Big Data. 2021. https://doi.org/10.1186/s40537021004680.
Shabbir MQ, Gardezi SBW. Application of big data analytics and organizational performance: the mediating role of knowledge management practices. J Big Data. 2020. https://doi.org/10.1186/s40537020003176.
Hu KH, Hsu MF, Chen FH, Liu MZ. Identifying the key factors of subsidiary supervision and management using an innovative hybrid architecture in a big data environment. Financ Innov. 2021. https://doi.org/10.1186/s40854020002199.
Shah SIH, Peristeras V, Magnisalis I. DaLiF: a data lifecycle framework for datadriven governments. J Big Data. 2021. https://doi.org/10.1186/s40537021004813.
Caraka RE, Chen RC, Huang SW, Chiou SY, Gio PU, Pardamean B. Big data ordination towards intensive care event count cases using fast computing GLLVMS. BMC Med Res Methodol. 2022. https://doi.org/10.1186/s12874022015384.
Daki H, El Hannani A, Aqqal A, Haidine A, Dahbi A. Big Data management in smart grid: concepts, requirements and implementation. J Big Data. 2017. https://doi.org/10.1186/s405370170070y.
Colombo P, Ferrari E. Access control technologies for big data management systems: literature review and future trends. Cybersecurity. 2019. https://doi.org/10.1186/s4240001800209.
Dash S, Shakyawar SK, Sharma M, Kaushik S. Big data in healthcare: management, analysis and future prospects. J Big Data. 2019. https://doi.org/10.1186/s4053701902170.
Lee Y, Nelder J. Modelling and analysing correlated nonnormal data. Stat Model. 2001;1:3–16.
Lee D, Lee W, Lee Y, Pawitan Y. Sparse partial leastsquares regression and its applications to highthroughput data analysis. Chemom Intell Lab Syst. 2011;109:1–8.
Lee Y, Nelder JA. Hierarchical generalized linear models. J Royal Stat Soc Ser B. 1996. https://doi.org/10.1111/j.25176161.1996.tb02105.x.
Hudak Jr SJ, Saxena A, Bucci RJ, Malcolm RC. Development of standard methods of testing and analyzing fatigue crack growth rate data. 1978.
Lu CJ, Meeker WO. Using degradation measures to estimate a timetofailure distribution. Technometrics. 1993;35:161–74.
Lee Y, Molas M, Noh M. mdhglm: multivariate double hierarchical generalized linear models. 2018.
Price CJ, Kimmel CA, Tyl RW, Marr MC. The developmental toxicity of ethylene glycol in rats and mice. Toxicol Appl Pharmacol. 1985;81:113–27.
Ha ID, Christian NJ, Jeong JH, Park J, Lee Y. Analysis of clustered competing risks data using subdistribution hazard models with multivariate frailties. Stat Methods Med Res. 2016;25:2488–505.
Ha ID, Lee Y, Song JK. Hierarchicallikelihood approach for mixed linear models with censored data. Lifetime Data Anal. 2002;8:163–76.
Hao L, Kim J, Kwon S, do Ha I. Deep learningbased survival analysis for highdimensional survival data. Mathematics. 2021;9:1–18.
Kim JM, do Ha I. Deep learningbased residual control chart for binary response. Symmetry. 2021;13:1–15.
Ha ID, Youngjo L. Multilevel mixed linear models for survival data. Lifetime Data Anal. 2005;11:131–42.
Lee Y, Ha ID. Orthodox BLUP versus hlikelihood methods for inferences about random effects in Tweedie mixed models. Stat Comput. 2010;20:295–303.
Ha ID, Lee Y. Estimating frailty models via poisson hierarchical generalized linear models. J Comput Graph Stat. 2003. https://doi.org/10.1198/1061860032256.
Crowder M, Fleming TR, Harrington DP. Counting processes and survival analysis. J Royal Stat Soc Ser A. 1994. https://doi.org/10.2307/2983370.
Fleming TR, Lin DY. Survival analysis in clinical trials: past developments and future directions. Biometrics. 2000. https://doi.org/10.1111/j.0006341X.2000.0971.x.
Cortes C, Vapnik V. Supportvector networks. Mach Learn. 1995;20:273–97.
Drucker H, Burges CJC, Kaufman L, Smola A, Vapnik V. Support vector regression machines. Adv Neural Inf Process Syst. 1996;9:155–61.
Williams CKI. Learning with kernels: support vector machines, regularization, optimization, and beyond. J Am Stat Assoc. 2003. https://doi.org/10.1198/jasa.2003.s269.
Fradkin D, Muchnik I. Support vector machines for classification. DIMACS series in discrete mathematics and theoretical computer science. Citeseer. 2006;70:13–20.
Schölkopf B. The kernel trick for distances. Adv Neural Inform Process Syst. 2001;13:301–7.
Wang J, Lee J, Zhang C. Kernel trick embedded Gaussian mixture model. Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics). 2003;2842:159–74.
Hofmann M. Support vector machines—kernels and the kernel trick. Universität Bamberg. 2006;26(3):1–16. http://www.cogsys.wiai.unibamberg.de/teachingarchive/ss06/hs_svm/slides/SVM_Seminarbericht_Hofmann.pdf.
Caraka RE, Nugroho NT, Tai SK, Chen RC, Toni T, Bens P. Feature importance of the aortic anatomy on endovascular aneurysm repair (EVAR) using Boruta and Bayesian MCMC. Commun Math Biol Neurosci 2020.
Caraka RE, Noh M, Lee Y. Package ‘hsem’. R project; 2021. p. 1–7.
Moutarde F. Deeplearning: general principles + convolutional neural networks. 2018
Czum JM. Dive into deep learning. J Am Coll Radiol. 2020. https://doi.org/10.1016/j.jacr.2020.02.005.
Wilson AG, Hu Z, Salakhutdinov R, Xing EP. Deep Kernel learning. Artificial intelligence and statistics (AISTATS). 2016;370378. http://arxiv.org/abs/1511.02222.
Benuwa BB, Zhan YZ, Ghansah B, Wornyo DK, Banaseka KF. A review of deep machine learning. Int J Eng Res Africa. 2016;24:124–36.
Schmidhuber J. Deep learning in neural networks: an overview. Neural Netw. 2015;61:85–117.
Sakti AD, Rinasti AN, Agustina E, Diastomo H, Muhammad F, Anna Z, et al. Multiscenario model of plastic waste accumulation potential in indonesia using integrated remote sensing, statistic and sociodemographic data. ISPRS Int J GeoInform. 2021. https://doi.org/10.3390/ijgi10070481.
Syahid LN, Sakti AD, Virtriana R, Wikantika K, Windupranata W, Tsuyuki S, et al. Determining optimal location for mangrove planting using remote sensing and climate model projection in southeast Asia. Remote Sens. 2020;12:1–29.
Sakti AD, Fauzi AI, Takeuchi W, Pradhan B, Yarime M, VegaGarcia C, et al. Spatial prioritization for wildfire mitigation by integrating heterogeneous spatial data: a new multidimensional approach for tropical rainforests. Remote Sens. 2022;14:543.
Sakti AD, Fauzi AI, Wilwatikta FN, Rajagukguk YS, Sudhana SA, Yayusman LF, et al. Multisource remote sensing data product analysis: investigating anthropogenic and naturogenic impacts on mangroves in southeast asia. Remote Sens. 2020;12:1–29.
Sakti AD, Rahadianto MAE, Pradhan B, Muhammad HN, Andani IGA, Sarli PW, et al. School location analysis by integrating the accessibility, natural and biological hazards to support equal access to education. ISPRS Int J GeoInform. 2022. https://doi.org/10.3390/ijgi11010012.
Hippert HS, Bunn DW, Souza RC. Large neural networks for electricity load forecasting: are they overfitted? Int J Forecast. 2005;21:425–34.
Zhang G, Patuwo BE, Hu MY. Forecasting with artificial neural networks: the state of the art. Int J Forecast. 1998;14:35–62.
Leung MT, Chen AS, Daouk H. Forecasting exchange rates using general regression neural networks. Comput Oper Res. 2000;27:1093–110.
Herliansyah R, Jamilatuzzahro J. Feed forward neural networks for forecasting indonesia exchange composite index. GSTF J Math Stat Oper Res. 2017. https://doi.org/10.5176/225133884.1.77.
Toharudin T, Pontoh RS, Caraka RE, Zahroh S, Lee Y, Chen RC. Employing long shortterm memory and facebook prophet model in air temperature forecasting. Commun Stat Simulat Comput. 2021;early acces:1–12.
Pontoh RS, Solichatus Z, Hidayat Y, Aldella R, Jiwani NM, Sukono. Covid19 modelling in south korea using a time series approach. Int J Adv Sci Technol. 2020;29:1620–32.
Lee Y, Nelder JA, Noh M. Hlikelihood: problems and solutions. Stat Comput. 2007;17:49–55.
Livieris IE, Pintelas E, Pintelas P. A CNN–LSTM model for gold price timeseries forecasting. Neural Comput Appl. 2020;32:17351–60. https://doi.org/10.1007/s0052102004867x.
Brodersen KH, Gallusser F, Koehler J, Remy N, Scott SL. Inferring causal impact using bayesian structural timeseries models. Ann Appl Stat. 2015;9:247–74.
Khodabakhsh AA. Forecasting multivariate timeseries data using LSTM and minibatches in the 7th international conference on contemporary issues in data science. Cham: Springer; 2019. p. 121–9.
Makridakis S, Spiliotis E, Assimakopoulos V. M5 accuracy competition: results, findings, and conclusions. Int J Forecast. 2022. https://doi.org/10.1016/j.ijforecast.2021.11.013.
Makridakis S, Hibon M. The M3competition: results, conclusions and implications. Int J Forecast. 2000. https://doi.org/10.1016/S01692070(00)000571.
Theodorou E, Wang S, Kang Y, Spiliotis E, Makridakis S, Assimakopoulos V. Exploring the representativeness of the M5 competition data. Int J Forecast. 2021. https://doi.org/10.1016/j.ijforecast.2021.07.006.
Makridakis S, Spiliotis E, Assimakopoulos V. The M4 Competition: 100,000 time series and 61 forecasting methods. Int J Forecast. 2020. https://doi.org/10.1016/j.ijforecast.2019.04.014.
Caraka RE, Chen RC, Yasin H, Pardamean B, Toharudin T, Wu SH. Prediction of status particulate matter 25 using state Markov chain stochastic process and HYBRID VARNNPSO. IEEE Access. 2019;7:161654–65.
Caraka RE, Chen RC, Yasin H, Lee Y, Pardamean B. Hybrid vector autoregression feedforward neural network with genetic algorithm model for forecasting spacetime pollution data. Indonesian J Sci Technol. 2021;6:243–66.
Aswale VA, Shaikh JA. Detection of microaneurysm in fundus retinal images using SVM classifier. IJEDR. 2017;5:175–80.
Pardamean B, Cenggoro TW, Rahutomo R, Budiarto A, Karuppiah EK. Transfer learning from chest Xray pretrained convolutional neural network for learning mammogram data. Proc Comput Sci. 2018;135:400–7. https://doi.org/10.1016/j.procs.2018.08.190.
Novitasari DCR, Hendradi R, Caraka RE, Rachmawati Y, Fanani NZ, Syarifudin A, et al. Detection of covid19 chest Xray using support vector machine and convolutional neural network. Commun Math Biol Neurosci. 2020.
Whi W, Ha S, Kang H, Lee DS. Hyperbolic disc embedding of functional human brain connectomes using resting state fMRI. bioRxiv. 2021. https://doi.org/10.1101/2021.03.25.436730.
Lee D, Kang H, Kim E, Lee H, Kim H, Kim YK, et al. Optimal likelihoodratio multiple testing with application to Alzheimer’s disease and questionable dementia data analysis, statistics and modelling. BMC Med Res Methodol. 2015;15:1–11.
Kim JY, Oh D, Sung K, Choi H, Paeng JC, Cheon GJ, et al. Visual interpretation of [18F]Florbetaben PET supported by deep learningbased estimation of amyloid burden. Eur J Nucl Med Mol Imag. 2021;48:1116–23.
Choi H, Ha S, Kang H, Lee H, Lee DS. Deep learning only by normal brain PET identify unheralded brain anomalies. EBioMedicine. 2019;43:447–53. https://doi.org/10.1016/j.ebiom.2019.04.022.
Whi W, Park JY, Choi H, Paeng JC, Cheon GJ, Kang KW, et al. Predicting outcome of repair of medial meniscus posterior root tear with early osteoarthritis using bone singlephoton emission computed tomography/computed tomography. Medicine. 2020;99: e21047.
Bae S, Choi H, Whi W, Paeng JC, Cheon GJ, Kang KW, et al. Spatial normalization using earlyphase [18F]FPCIT PET for quantification of striatal dopamine transporter binding. Nucl Med Mol Imag. 2020;54:305–14.
Whi W, Huh Y, Ha S, Lee H, Kang H, Lee DS. Characteristic functional cores revealed by hyperbolic disc embedding and kcore percolation on restingstate fMRI. Sci Rep. 2022. https://doi.org/10.1038/s41598022089757.
Funding
National Research Foundation of Korea fully supports this work under Grants [NRF2019R1A2C1002408].
Author information
Authors and Affiliations
Contributions
Conceptualization: REC, YL, JH, HL, MN, IDH. Methodology: REC, YL, JH, HL, MN, IDH. Project Administration: REC, YL, MN, IDH, BP. Software: REC, YL, JH, HL, MN, IDH, PUG. Validation: REC, YL, JH, HL, MN, IDH. Visualization: REC, YL, JH, HL, MN, IDH. Writing—original draft, review and editing: REC, YL, JH, HL, MN, IDH, PUG, and BP. All authors read and approved the final manuscript.
Corresponding authors
Ethics declarations
Competing interests
The authors declare that they have no known competing financial interests or personal relationships that could have influenced the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix A
Hlikelihood theory for the frailty model
The hlikelihood gives a straightforward way of handling nonparametric baseline hazards. The hlikelihood under the frailty model is defined by:
Here,
The functional form of \({\lambda }_{0}(t)\) is unknown. Hence, we consider \({\wedge }_{0}\)(t) to be a step function with jumps at the observed event time/At the moment, \({y}_{\left(k\right)}\) is the kth smallest distinct event time among the \({y}_{ij}\)’s, and \({\lambda }_{0k}\) = \({\lambda }_{0}\left({y}_{(k)}\right)\). Thus, we proposed the use of the profile hlikelihood with \({\lambda }_{0}\) eliminated, \({r}^{*}=h{}_{{\lambda }_{0}=\widehat{{\lambda }_{0}}},\) given by
Here,
where,
are solutions of the estimating equations, \(\frac{\partial h}{\partial {\lambda }_{0k}}=0\). However, \({d}_{\left(k\right)}\) is the number of events at \({y}_{\left(k\right)}\) and \({R}_{\left(k\right)}=\left\{{\left(i,j\right): y}_{ij}\ge {y}_{\left(k\right)}\right\}\) is the risk set at \({y}_{\left(k\right)}\). In consequences, we proposed hlikelihood, called penalized partial likelihood (PPL) \({h}_{p}\), given by:
Appendix B
Calculation of scaled deviance test
The scaled deviance is defined by following equation:
Here, the estimated degree of freedom, \(d.f=ntr\left({H}^{1}{H}^{*}\right),\) where \(\mathrm{H}\) and \({H}^{*}\) are the Hessian matrices of \((\beta ,v)\) based on \({l}_{0}\) and \(\mathrm{h},\) respectively.
Appendix C
Conditional Akaike information criteria towards DHGLM
The conditional Akaike information for double HGLMs is defined as follows:
Here \(f\left(y,u,a\right)={f}_{1}\left(yu,a\right){f}_{2}(u){f}_{3}(a)\) is the true joint distribution of \(y,u,\) and\(a\). Meanwhile, \(\widehat{\beta }\left(y\right) \mathrm{and} \widehat{v}\left(y\right)\) are the estimators of fixed and random effects (\(\beta ,\upnu\)) for the mean model, respectively. Here, \(\widehat{\gamma }\left(y\right)\) and \(\widehat{b}\left(y\right)\) are also the estimators of fixed and random effects\((\gamma ,b\)) for the dispersion model, respectively. At the same time, another two evaluation criteria are mAIC for marginal loglikelihood and rAIC for restricted loglikelihood [5], defined by:
Here \(d{f}_{m}\) is the number of fixed parameters and \(d{f}_{r}\) is the number of dispersion parameters. When we compare models with different fixed parameters, mAIC can be used, whereas rAIC can be used for dispersion parameter model selection.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Caraka, R.E., Lee, Y., Han, J. et al. Albatross analytics a handson into practice: statistical and data science application. J Big Data 9, 70 (2022). https://doi.org/10.1186/s4053702200626y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s4053702200626y