Keywords

1 Introduction

Suppose we have random samples from two logistic populations with a common scale parameter \(\sigma \) and possibly distinct and unknown location parameters \(\mu _{1}\) and \(\mu _{2}.\) Specifically, let and be m and n independent random samples taken from two logistic populations Logistic\((\mu _{1},\sigma )\) and Logistic\((\mu _{2},\sigma )\) respectively. Here we denote Logistic\((\mu _{i}, \sigma );\) \(i=1,2\) as the logistic population having the probability density function

$$\begin{aligned} f(x;\mu _{i},\sigma )=\frac{\frac{1}{\sigma }\exp \{-(\frac{x-\mu _{i}}{\sigma })\}}{[1+\exp \{-(\frac{x-\mu _{i}}{\sigma })\}]^2},~-\infty<x<\infty ,~-\infty<\mu _{i} <\infty ,~ \sigma > 0 \end{aligned}$$
(1.1)

The problem is to classify a new observation into one of the two logistic populations using some of the estimators of the common scale parameter and two different location parameters. First, we will derive certain estimators of the associated parameters such as \(\sigma ,\) and the other two nuisance parameters \(\mu _{1}\) and \(\mu _{2}\) from a Bayesian perspective. Using these estimators, we will construct some classification rules to classify a new observation into one of the two logistic populations using training samples. The performances of all the estimators will be evaluated through their bias

$$\begin{aligned} E(d-\theta ) \end{aligned}$$
(1.2)

and the mean squared error (MSE)

$$\begin{aligned} E(d-\theta )^2, \end{aligned}$$
(1.3)

where ‘d’ is an estimator for estimating the parameter \(\theta .\)

In a general framework, the problem of classification is to classify a new observation, say z into one of the several populations/groups, say \(\pi _1,\) \(\pi _2,\) \(\ldots ,\) \(\pi _k,\) using certain methodologies and training data. The classification problem can be seen in machine learning for image classification, text classification, electrical science for signal classification, military surveillance for recognizing speech, in biomedical sciences for taxonomic study. It also has an application in computer science for classifying the object using receiver operating characteristic (ROC) curve.

Researchers in the past have considered the problem of classification under the equality assumption of parameters. For example, [12] considered the classification problem under equality assumption on the location parameters using training samples from several shifted exponential populations. Utilizing some well-known estimators of the common location parameter, the authors proposed several classification rules. The performances of all the rules have been evaluated using the expected probability of misclassification. Further, [13] considered the classification rules for the two Gaussian populations with a common mean. It is noted that the classification problem was first considered by Fisher in the year 1936 in a general framework. However, under multivariate normal model, the problem of classification was considered by [1, 2, 16]. Basu [5] proposed the classification rules for the two shifted exponential distribution. Note that [2] considered the classification problem under multivariate normal set up with a common co-variance matrix. This model reduces the problem under common standard deviation (scale) in the case of univariate normal.

Contrary to the above, significantly less attention has been paid to constructing classification rules and studying their performances when the distribution is not normal and exponential. However, there are many practical scenarios where the emerging data sets satisfactorily modeled by logistic distribution. In view of this, we take two logistic populations having a common scale parameter (which is also the standard error) and two different location parameters and construct several classification rules. In this regard, we have considered two real-life data sets from civil engineering science, and the other is related to food science, which has been satisfactorily modeled by using logistic distribution. The details of which have been discussed in Sect. 5. Using the estimators under the same model problem, we construct specific classification rules that may classify a new observation. The applications of logistic distribution are seen in several fields of study. Logistic distribution is use to describe the modeling agriculture production data, growth model in a demographic study (see [4]), analysis of biomedical data, life testing data and bio-assay (see [14, 21]).

Several authors have considered the problem of estimating parameters under equality restrictions in the past. For recent detailed and updates literature on estimating parameters under equality restriction under the various probabilistic model, we refer to [18,19,20, 23,24,25,26]. Recently, [20] studied the estimation of common scale parameter (\(\sigma \)) of two logistic populations with the unknown and unequal location parameters (\(\mu _{i}\)) from Bayesian point of view. They derived the MLEs and individual approximate Bayes estimators of the parameter using non-informative and informative priors, and using the Monte Carlo simulation study; they compare the performances of all the proposed Bayes estimators numerically.

The researchers studied the point and interval estimation of logistic population parameters, namely the location and scale parameters, from classical and Bayesian points of view. The first attempt to estimate the logistic distribution parameters was made by [9], where they derived the linear unbiased estimators based on sample quantiles and some approximate unbiased linear estimators and compared their efficiencies numerically. One may refer to [3, 7, 11, 17] for results on estimating scale and location parameters of a logistic distribution.

The remaining contribution of the current research work can be described as follows. In Sect. 2, we discuss the Bayes classification rule in a general framework and then obtained the classification rule for our proposed model set up. Consequently, using the maximum likelihood estimators (MLEs) of the associated parameters, we construct a classification rule. In Sect. 3, we consider Bayes estimators of the associated parameters using Markov chain Monte Carlo methods. Using these estimators, classification rule also has been constructed. In Sect. 4, we carry out a detailed simulation study and compute the bias and the mean squared error of the proposed estimators. Moreover, these estimators are compared with some of the existing estimators previously proposed by Nagamani [20] for the associated parameters. Further classification rules are constructed using the estimators proposed by Nagamani [20] and compared with the rules using MLEs and the MCMC method. In Sect. 5, we present two real-life data sets, which has been satisfactorily modeled by logistic distributions with common scale parameter. A new observation has been classified into one of the data sets, using the proposed classification rules. Concluding remarks are given in Sect. 6.

2 Classification Rule Using the MLE for Two Logistic Populations

In this section, we construct a classification rule using the MLEs of the associated parameters, \(\sigma ,\) \(\mu _{1}\) and \(\mu _{2}\) to the model (1.1).

Suppose we have training samples and from two Logistic populations \(\varPi _{1}\) and \(\varPi _{2}\) respectively, where \(\varPi _{i}\) is the population having density \(f_{i}(x;\mu _i,\sigma ),\) \(i=1,2.\) Now a new observation, say z is classified into the population \(\varPi _{1}\) if

$$\begin{aligned} \log \bigg ( \frac{f_{1}({z})}{f_{2}({z})}\bigg ) \ge \log \bigg (\frac{C(2|1)*q_{2}}{ C(1|2)*q_{1} }\bigg ) \end{aligned}$$
(2.1)

and into the population \(\varPi _{2}\) if

$$\begin{aligned} \log \bigg (\frac{f_{1}({z})}{f_{2}({z})}\bigg ) < \log \bigg (\frac{C(2|1)*q_{2}}{ C(1|2)*q_{1} }\bigg ). \end{aligned}$$
(2.2)

Here C(i|j) is the cost of misclassification when an observation is from \(\varPi _j\) and is misclassified into the population \(\varPi _i,\) \(i\ne j,\) \(i,j=1,2.\) If we assume that \(C(2|1)=C(1|2)\) and \(q_{2}= q_{1},\) where \(q_{i}\) is the prior probability associated to the population \(\varPi _{i}\) then, the classification function for classifying an observation z into Logistic\((\mu _{1},\sigma )\) or Logistic\((\mu _{2},\sigma )\) obtained as

$$\begin{aligned} W(z)= & {} \frac{1}{{\sigma }}\bigg ((z-{\mu }_{2}) -(z-{\mu }_{1})\bigg ) + 2\log \bigg (1+\exp \bigg (-\frac{(z-{\mu }_{2})}{{\sigma }}\bigg ) \bigg )\nonumber \\&-2\log \bigg (1+\exp \bigg (-\frac{(z-{\mu }_{1})}{{\sigma }}\bigg ) \bigg ). \end{aligned}$$
(2.3)

When the parameters are unknown, a natural approach is to replace the parameters by their respective estimators and obtain the classification rule. Thus utilizing the classification function W(z),  we propose classification Rule R as: classify z into the population \(\varPi _{1}\) if \(W(z)\ge 0,\) else classify z into the population \(\varPi _{2}.\)

Let us denote P(i|jR),  (\(i\ne j,\) \(i,j=1,2\)) as the probability of misclassification for an observation from \(\varPi _{j}\) when it is misclassified into the population \(\varPi _{i}.\) The new observation z is classified into one of the two populations, such that the probability of misclassification becomes the minimum. We refer to Anderson [1], for the details of the classification rules using prior information.

In order to obtain the MLEs of the associated parameters, let us consider the log-likelihood function under the current model setup which is given by

(2.4)

where \(-\infty<\mu _i<\infty (i=1,2);\) \(\sigma >0.\)

Note that, under the same model setup the MLEs of the parameters \(\sigma ,\) \(\mu _{1}\) and \(\mu _{2}\) have been numerically obtained by Nagamani [20]. Let us denote the MLEs of the parameters \(\sigma ,\) \(\mu _{1}\) and \(\mu _{2}\) respectively by \(\hat{\sigma }_{ml},\) \(\hat{\mu }_{1ml}\) and \(\hat{\mu }_{2ml}.\) For computing the MLEs, using Newton–Raphson method, we have taken the initial guess as suggested by Nagamani [20].

Now using the MLEs of the parameters \(\mu _{1},\) \(\mu _{2}\) and \(\sigma ,\) we define a classification function, say \(W_{ML}(z)\) as,

$$\begin{aligned} W_{ML}(z)= & {} \frac{1}{\hat{\sigma }_{ml}}\bigg ((z-\hat{\mu }_{2ml}) -(z-\hat{\mu }_{1ml})\bigg ) + 2\log \bigg (1+\exp \bigg (-\frac{(z-\hat{\mu }_{2ml})}{\hat{\sigma }_{ml}}\bigg ) \bigg )\nonumber \\&-2\log \bigg (1+\exp \bigg (-\frac{(z-\hat{\mu }_{1ml})}{\hat{\sigma }_{ml}}\bigg ) \bigg ). \end{aligned}$$
(2.5)

Using this classification function \(W_{ML}(z),\) we propose a classification rule, say \(R_{ML}\) as: classify z into \(\varPi _{1}\) if \(W_{ML}(z) \ge 0,\) else classify z into the \(\varPi _{2}.\)

3 Bayesian Estimation of Parameters Using MCMC Approach and Classification Rule

It is noted that certain Bayes estimators of parameters using Lindley’s approximation method and different types of priors with respect to the squared error loss function and the LINUX loss function have been obtained by Nagamani [20]. Here, we consider Bayes estimators of the associated parameters using MCMC method along with the Metropolis–Hastings algorithm. Moreover, utilizing these estimators we propose a classification rule. Metropolis–Hastings algorithm is an important class of MCMC method, which is applied to get the samples from the posterior distribution effectively in a systematic manner.

A special case of Metropolis–Hastings method is the symmetric random walk Metropolis (RWM) algorithm. In this method, we take the Markov chain described by c alone to be a simple symmetric random walk, so that \(c(\delta , \delta ^{*}) =c(\delta ^{*}-\delta )\) that is the probability density c is a symmetric function. Suppose that a target distribution has probability density function \(\pi .\) Thus for given \(\delta _n,\) (at \(n^{th}\) state) we generate \(\delta ^{*}_{n+1}\) from a specified probability density, say \(q(\delta _{n},\delta ^{*})\) and is then accepted with probability \(\alpha (\delta _{n}, \delta ^{*}_{n+1})\) which is given by

$$\begin{aligned} \alpha (\delta ,\delta ^{*}) = \min \bigg (\frac{\pi (\delta ^{*})}{\pi (\delta )} , 1\bigg ). \end{aligned}$$
(3.1)

If the proposed value is accepted, we set \(\delta _{n+1} = \delta ^{*}_{n+1},\) otherwise we reject it and assign \(\delta _{n+1}=\delta _{n}.\) The function \(\alpha (\delta , \delta ^{*})\) is chosen precisely to ensure that the Markov chain \(\delta _0, \delta _1, \dotsc , \delta _n\) is reversible with respect to the target probability density \(\pi (\delta ^{*}),\) in the sense that the target density is stationary for the chain. We refer to [6, 8, 10] for details of the procedure of the Metropolis–Hastings algorithm.

In order to estimate the parameters using Metropolis–Hastings algorithm, in the current model, we need to specify the prior distributions for the unknown parameters \(\sigma ,\) \(\mu _{1}\) and \(\mu _{2}.\) We assume that the three parameters are independent having the prior probability densities respectively as

$$\begin{aligned} \mu _i \sim N(a_{i}, b_{i}^{2}), ~ \sigma \sim \varGamma (c,d); i=1,2, \end{aligned}$$
(3.2)

where \(N(a_{i}, b_{i}^{2})\) denotes the normal distribution having mean \(a_{i}\) and variance \(b_{i}^{2}\) and \(\varGamma (c,d)\) is the gamma distribution having c as shape and d as scale parameters, respectively. These hyperparameters are known. The posterior probability density of the parameters is seen to be of the following forms.

where we denote \(g(\mu _{1},\sigma _{1},x_{i}){=}\exp \{-(\frac{x_{i}-\mu _{1}}{\sigma })\},\) \(h(\mu _{2},\sigma ,y_{j})=\exp \{-(\frac{y_{j}-\mu _{2}}{\sigma })\},\) \(K_{1}=\sum _{i=1}^{m}(x_{i}-\mu ),\) and \(K_{2}=\sum _{j=1}^{n}(y_{j}-\mu ).\) Note that, the posterior densities of \(\mu _{1},\) \(\mu _{2}\) and \(\sigma \) do not have any known distributional form. Hence, to generate samples from the posterior probability distributions of \(\mu _{1},\) \(\mu _{2}\) and \(\sigma ,\) we use the well-known random walk Metropolis–Hastings algorithm (RWM) which is a MCMC procedure. The details of the computational steps to generate \(\mu _{1},\) \(\mu _{2}\) and \(\sigma \) simultaneously using Metropolis–Hastings algorithm can be described as follows.  

Step-1::

Let the \(k^{th}\) state of Markov chain consists of \((\mu _{1}^{(k-1)}, \mu _{2}^{(k-1)}, \sigma ^{(k-1)}).\)

Step-2::

In order to update the parameters using random walk Metropolis–Hastings algorithm (RWM), we generate \(\varepsilon _{1} \sim N(0,\sigma _{\mu _{1}}^{2}),\) \(\varepsilon _{2} \sim N(0,\sigma _{\mu _{2}}^{2})\) and \(\varepsilon _{3} \sim N(0,\sigma _{\sigma }^{2}).\) Set \(\mu _{1}^{(*)}=\mu _{1}^{(k-1)}+\varepsilon _{1},\) \(\mu _{2}^{(*)}=\mu _{2}^{(k-1)}+\varepsilon _{2},\) \(\sigma ^{(*)}=\sigma ^{(k-1)}+\varepsilon _{3}.\)

Step-3::

Calculate the term \(\alpha (\mu _{1}^{(*)},\mu _{2}^{(*)},\sigma ^{(*)})\) as given (3.1).

Step-4::

Define a term \(S=\min (1, \alpha (\mu _{1}^{(*)},\mu _{2}^{(*)},\) \(\sigma ^{(*)})).\) Generate a random number \(v \sim U(0,1).\) If \( v \le S,\) accept \((\mu _{1}^{(*)},\mu _{2}^{(*)},\) \(\sigma ^{*})\) and update the parameters as \(\mu _{1}^{(k)}= \mu _{1}^{(*)},\) \(\mu _{2}^{(k)}=\mu _{2}^{(*)}\) and \(\sigma ^{(k)}=\sigma ^{(*)},\) otherwise reject \(\mu _{1}^{(*)},\) \(\mu _{2}^{(*)}\) and \(\sigma ^{(*)}\) and set \(\mu _{1}^{(k)}=\mu _{1}^{(k-1)},\) \(\mu _{2}^{(k)}=\mu _{2}^{(k-1)}\) and \(\sigma ^{(k)}=\sigma ^{(k-1)}.\) Repeat this step for \(k=1,2,\ldots , K,\) where K is a large number suitably fixed.

 

In this computational method, the most important step is to choose the values of \(\sigma _{\mu _{1}}^{2},\) \(\sigma _{\mu _{2}}^{2}\) and \(\sigma _{\sigma }^{2}.\) One may refer to [6] regarding choice of these variances. The values of \(\sigma _{\mu _{1}}^{2},\) \(\sigma _{\mu _{2}}^{2}\) and \(\sigma _{\sigma }^{2}\) are chosen in such a way that a acceptance ratio should lie within the range from 20% to 30%. Now using these MCMC samples, we estimate the parameters \(\mu ,\) \(\sigma _{1}\) and \(\sigma _{2}\) respectively as

$$\begin{aligned} \hat{\mu }_{1mc}=\frac{1}{M-M_0}\sum _{i=M_0+1}^{M}\mu _{1}^{(i)}, ~\hat{\mu }_{2mc}=\frac{1}{M-M_0}\sum _{i=M_0+1}^{M}\mu _{2}^{(i)} ~\text{ and } ~\hat{\sigma }=\frac{1}{M-M_0}\sum _{i=M_0+1}^{M}\sigma ^{(i)}, \end{aligned}$$
(3.3)

where \(M_{0}\) is the burn-in period for the RWM method.

Utilizing these Bayes estimators of the parameters \(\mu _{1},\) \(\mu _{2}\) and \(\sigma ,\) we construct a classification function, say \(W_{MC}(z)\) as

$$\begin{aligned} W_{MC}(z)= & {} \frac{1}{\hat{\sigma }_{}}\bigg ((z-\hat{\mu }_{2mc}) -(z-\hat{\mu }_{1mc})\bigg ) + 2\log \bigg (1+\exp \bigg (-\frac{(z-\hat{\mu }_{2mc})}{\hat{\sigma }_{mc}}\bigg ) \bigg )\nonumber \\&-2\log \bigg (1+\exp \bigg (-\frac{(z-\hat{\mu }_{1mc})}{\hat{\sigma }_{mc}}\bigg ) \bigg ). \end{aligned}$$
(3.4)

Using this classification function \(W_{MC}(z),\) we propose a classification rule, say \(R_{MC}\) as: classify z into \(\varPi _{1}\) if \(W_{MC}(z) \ge 0,\) else classify z into the \(\varPi _{2}.\)

Remark 1

We note that, all the existing estimators such as the MLE, Bayes estimators using different priors for the current model are compared with the Bayes estimator using MCMC approach in terms of MSE and bias numerically in Sect. 4. It has been noticed that the MLEs and the Bayes estimators using MCMC approach outperform the earlier estimators for all sample sizes. Hence, in the numerical comparison section, we only consider classification rules for comparison purpose based on the MLEs and the Bayes estimators using the MCMC approach.

4 Numerical Comparison of Classification Rules

In this section, we will evaluate and compare the performances of all the classification rules in terms of probability of misclassification, numerically. Note that none of the estimators, and hence the classification rules, could be obtained in closed form expressions. It is not possible to compare their performances analytically. However, utilizing the advanced computational facilities available nowadays, we have compared the performances of all the proposed rules numerically, which will be handy in certain practical applications.

In Sect. 2, we have proposed the classification rule \(R_{ML}\) using the MLEs of the associated parameters for the two logistic populations. In Sect. 3, we have proposed the classification rule \({R}_{MC}\) utilizing the Bayes estimators through MCMC approach. In our simulation study, we have also included estimators proposed by Nagamani [20] and compared their MSE and bias with the Bayes estimator that uses MCMC approach (our proposed estimator). It has been noticed that the MLEs as well as the proposed Bayes estimator through MCMC approach always have minimum MSE and bias. Hence, we have not considered the classification rules based on the Bayes estimators proposed by Nagamani [20], for numerical comparison in terms of probability of misclassification.

In order to compute the probability of misclassification of the classification rules \(R_{ML}\) and \({R}_{MC},\) we proceed in the following manner.  

Step-1::

Generate training samples of size m from the logistic population Logistic\((\mu _{1},\sigma ).\) Then, similarly, we generate training samples of size n from another logistic population Logistic\((\mu _{2},\sigma ).\) Utilizing these training samples, we estimate the parameters \(\mu _{1},\) \(\mu _{2}\) and \(\sigma .\) In particular, we compute the MLEs and the Bayes estimators using MCMC approach that is we compute \(\hat{\mu }_{1ml},\) \(\hat{\mu }_{2ml},\) \(\hat{\sigma }_{ml},\) \(\hat{\mu }_{1mc},\) \(\hat{\mu }_{2mc},\) \(\hat{\sigma }_{mc},\) which are involved in the classification functions \(W_{ML}(z)\) and \(W_{MC}(z).\)

Step-2::

Generate a new observation, say z from the logistic population Logistic\((\mu _{1},\sigma )\) and substitute it in the classification rules \(R_{ML}\) and \(R_{MC},\) then we check whether it belongs to the population \(\varPi _{1}\) or not. Similarly, we generate a new sample z from the logistic population \(\varPi _{2},\) and substitute it in the classification rules, then check whether it belongs to the population \(\varPi _{2}\) or not.

  The above procedure is carried out using the well-known Monte Carlo simulation method with 20, 000 replications. A high level of accuracy has been achieved in the sense that the standard error in simulation is of the order of \(10^{-4}.\) . The probability of misclassification P(1|2) and P(2|1) using the probability frequency approach has been computed. In the computation of Bayes estimator using MCMC approach, the hyperparameters involved in the conjugate prior have been suitably taken as \(c=2.5,\) \(d=1\) (for the common scale parameter), \(a_1=2.5,\) \(b_1=3,\) \(a_2=2,\) \(b_2=1\) (in the case of location parameters). The computation also has been done using some other choices of hyperparameters; however, the probability of misclassification of \(R_{MC}\) changes insignificantly. In the MCMC method that uses Metropolis–Hastings algorithm, we have generated 50, 000 MCMC samples from the posterior density of \(\sigma ,\) \(\mu _{1}\) and \(\mu _{2}\) with burn-in period \(M_{0}=5000.\) In the simulation study, we have seen that for some rules, the P(1|2) may be small but P(2|1) may be high. In view of this and to have a better picture of the performance, we also define the expected probability of misclassification (EPM), given by

$$\begin{aligned} E(R)=\frac{P(1|2,R)+ P(2|1,R)}{2} \end{aligned}$$

along with the probability of misclassifications, for comparing the rules. We consider the cost and the prior probabilities are equal, that is \(C(1|2)=C(2|1)\) and \(q_1=q_2,\) for convenience.

Table 1 Expected probability of misclassification (EPM) for the proposed rules with \(\mu _1=1\) and \(\mu _2=2\)  and Various Values of \(\sigma \)

The simulation study has been carried out by considering various combinations of sample sizes and different choices of parameters. However, for illustration purpose we present the expected probability of misclassification of the rules \(R_{ML}\) and \({R}_{MC}\) for some specific choices of sample sizes and parameters. In Table 1, we have presented the EPM of the rules \(R_{ML}\) and \(R_{MC}\) for equal and unequal sample sizes, for fixed values of \(\mu _{1}\) and \(\mu _{2}\) with variation in scale parameter \(\sigma .\) The following observations were made from our simulation study as well as the Table 1, regarding the performances of the classification rules.

  • The expected probability of misclassification for the proposed rules decrease as the sample sizes increase. It is also noticed that for different location points (\(\mu _{i}\)) and fixing the scale parameter \(\sigma ,\) the EPM remains almost same. It seems that the rules are location invariant.

  • The rule \({R}_{MC}\) based on the MCMC procedure always performs better than the rule \(R_{ML}\) based on MLE in terms of expected probability of misclassification. It is further noticed that when \(\sigma \) is close to 4,  the maximum values for EPM is 0.5 for both the rules.

  • Considering the EPM as measure of performance, the classification rule \({R}_{MC}\) has the lowest EPM value among all the proposed rules. Hence, we recommend to use the classification rule \({R}_{MC}\) for classifying a new observation into one of the two logistic populations.

5 Application with Real-Life Data Sets

In this section, we consider two real-life data sets which are related to compressive strength of bricks and dietary fiber content in food. It has been shown that the logistic distribution is a reasonable fit to these data sets. Further, Levene’s test with significance level 0.05 indicates that the equality of scale parameters cannot be rejected (see [20]).

Example 1

Nagamani [20] considered the data sets related to compressive strength (MPa) of clay bricks and fly ash bricks. The experiment was conducted by Teja [22] to determine the mechanical properties, such as initial rate of absorption (IRA), water absorption (WA), dry density (DD), and compressive strength (CS) of 50 brick units from each type. The compressive strength (MPa) of clay bricks and fly ash bricks is given as follows.

Clay Brick: 8.02,  7.31,  7.31,  7.87,  7.09,  9.75,  5.90,  6.76,  6.58,  7.70,  5.70,  6.56,  8.14,  7.28,  5.70,  7.38,  7.02,  5.95,  5.90,  5.67,  7.22,  6.76,  7.73,  8.65,  8.08,  7.98,  5.60,  8.66,  9.67,  8.18,  8.70,  4.86,  9.33,  7.77,  6.21,  7.74,  11.31,  9.13,  8.28,  7.09,  5.62,  11.88,  5.73,  9.21,  7.03,  9.07,  7.81,  6.70,  9.97,  8.85

Fly Ash Brick: 3.62,  4.74,  9.88,  5.93,  6.09,  6.94,  6.32,  5.30,  5.14,  4.55,  4.03,  7.36,  3.57,  3.98,  4.03,  4.74,  7.32,  3.23,  5.38,  7.18,  6.07,  3.62,  6.64,  5.58,  5.23,  3.95,  5.86,  5.58,  6.97,  5.05,  4.35,  4.55,  4.79,  4.03,  4.74,  7.58,  3.62,  6.01,  3.99,  6.04,  4.74,  7.21,  3.61,  5.69,  7.21,  6.40,  3.55,  8.70,  4.35,  7.51.

It is our interest to classify a new observation, say z into Clay Brick or Fly Ash Brick. So, we compute the estimators for the parameters \((\mu _{1},\mu _{2},\sigma )\) as \((\hat{\mu }_{1ml},\hat{\mu }_{2ml},\hat{\sigma }_{ml})=(7.53694,5.35400,0.84857)\) and \((\hat{\mu }_{1mc},\hat{\mu }_{2mc},\hat{\sigma }_{mc})=(7.02195,4.86868,0.78183).\) Utilizing these estimators, we compute the classification functions \(W_{ML}(z)\) and \(W_{MC}(z).\)

Suppose we have a new observation, say \(z=6.5\) and we want to classify this observation into one of the two types of bricks using our proposed classification rules. The values of classification functions are computed as \(W_{ML}= 0.07285\) and \(W_{MC}=-0.91663.\) The rule \(R_{ML}\) classifies \(z=6.5\) into the clay brick, whereas the rule \(R_{MC}\) classifies it into the fly ash brick.

Example 2

These data sets are related to the percentage of total dietary fiber (TDF) content in foods, such as fruits, vegetables, and purified polysaccharides. Li and Cardozo [15] conducted an inter laboratory (nine laboratories involved) study using a procedure as described by Association of Official Agricultural Chemists (AOAC) international to determine the percentage of total dietary fiber (TDF) content in food samples. The experiment was conducted for six different types of food samples, such as (a) apples, (b) apricots, (c) cabbage, (d) carrots, (e) onions, and (f) FIBRIM 1450 (soy fiber). The percentage of TDF using non-enzymatic-gravimetric method from nine laboratories for apples and carrots is given as, Apple: 12.44,  12.87,  12.21,  12.82,  13.18,  12.31,  13.11,  14.29,  12.08;  Carrots: 29.71,  29.38,  31.26,  29.41,  30.11,  27.02,  30.06,  31.30,  28.37.

Nagamani [20] shown that logistic distribution fits these two data sets reasonably well. Their Levene’s test also confirms the equality of the scale parameters with level of significance 0.05. In order to classify an observation into these two types of foods, we compute the estimators of the parameters \((\mu _{1},\mu _{2},\sigma )\) as \((\hat{\mu }_{1ml},\hat{\mu }_{2ml},\hat{\sigma }_{ml})=(12.765,29.6497,0.5604)\) and \((\hat{\mu }_{1mc},\hat{\mu }_{2mc},\hat{\sigma }_{mc})=(12.65066,29.47925,0.63596).\) Utilizing these estimators, we compute the classification functions \(W_{ML}(z)\) and \(W_{MC}(z).\)

Suppose we have a new observation, say \(z=21.2\) and we want to classify this observation into one of the two types of foods using our proposed classification rules. The values of classification functions are computed as \(W_{ML}= 0.026588\) and \(W_{MC}=-0.314322.\) The rule \(R_{ML}\) classifies \(z=21.2\) into Apples, whereas the rule \(R_{MC}\) classifies it into the Carrots.

6 Concluding Remarks

In this note, we have considered the problem of classification into one of the two logistic populations with a common scale parameter and possibly different location parameters. It is worth mentioning that the same model was previously considered by Nagamani [20] and estimated the associated parameters. Specifically the authors had derived certain Bayes estimators using different types of priors and Lindley’s approximations. Utilizing their proposed estimators, and the one we have proposed (Bayes estimators using MCMC approach), we have constructed several classification rules. It has been seen that the proposed rule \(R_{MC}\) outperforms all other rules in terms of expected probability of misclassification. The application of our model problem has been explained using two real-life data sets.