亚洲免费av电影一区二区三区,日韩爱爱视频,51精品视频一区二区三区,91视频爱爱,日韩欧美在线播放视频,中文字幕少妇AV,亚洲电影中文字幕,久久久久亚洲av成人网址,久久综合视频网站,国产在线不卡免费播放

        ?

        Diagnostics in generalized nonlinear models based on maximum Lq-likelihood estimation

        2013-01-08 08:39:45XuWeijuanLinJinguan

        Xu Weijuan Lin Jinguan

        (Department of Mathematics, Southeast University, Nanjing 211189, China)

        Linear regression diagnostics have developed well in the past three decades. A comprehensive study on this topic can be found, for instance, in Refs.[1-2]. However, there has not been much published work on regression diagnostics for the models outside of linear regression[3-4]. Only a few papers are related to the diagnostics for exponential family nonlinear models[4-6]. Wei[7]described the diagnostics for generalized nonlinear models systematically in chapter 6.

        All the discussion above is based on the maximum likelihood method. Standard large sample theory guarantees that the maximum likelihood estimator (MLE) is asymptotically efficient. It means that when the sample is large, the MLE is at least as accurate as any other estimator. However, when the sample is moderate or small, the properties of the MLE may not be very good. So, we need a new estimation method, which can make the property better.

        In this paper, we propose a modified diagnostic method for a generalized nonlinear model based on the maximumLq-likelihood estimator[8]. Diagnostics for regression parameters and dispersion parameters are considered, and three kinds of diagnostic statistics are proposed. For small and modest sample sizes, the proposed diagnostic method is still available whenqis properly chosen. Some simulations are performed to investigate the behavior of the proposed methods for different sample sizes. We also show that the modified diagnostic statistics method outperform the classical diagnostic statistics method when the sample sizes are modest or even small.

        1 Generalized Entropy and Maximum Lq-Likeli-hood Estimator

        The Kullback-Leibler (KL) divergence[9]is one of the most popular quantities employed to measure the distance of a distribution with respect to a “true” distribution. Consider aσ-finite measureμon a measurable spaceΩand letΜbe the set of all probability distribution functionsf. The expectation with respect tofis denoted byEf. The KL divergence between two density functionsgandfis defined as

        (1)

        Note that finding the densitygthat minimizes Δ(f‖g) is equivalent to minimizing Shannon’s entropy[10]:

        Η(f‖g)=-Eflogg(X)

        (2)

        Definition1Letfandgbe two density functions; and theq-entropy is defined as

        Ηq(f,g)=-Ef[Lq(g(X))]q>0

        (3)

        where

        (4)

        The functionLqrepresents a Box-Cox transformation in statistics. In other fields it is often called a deformed logarithm. The above characterization emphasizes the similarity to the classical Shannon’s entropy: ifq→1, thenLq(u)→log(u) and the usual definition of Shannon’s entropy is recovered.

        Recently, Ferrari and Yang[8]introduced an estimator inspired by Havrda, and Charvt[11]generalized information measures (usually calledα-order entropies orq-entropies in physics), the maximumLq-likelihood estimator.

        Definition2LetX1, …,Xnbe an i.i.d. sample fromf(x;θ0),θ0∈Θ. The maximumLq-likelihood estimator ofθ0is defined as

        (5)

        whereLqis theq-logarithmic function defined in Eq.(4) withq> 0. TheLq-likelihood equation is

        (6)

        Note that when the distortion parameterqtends to 1,Lq(·)→log(·) and the usual MLE is recovered. In this sense, the maximumLq-likelihood estimation extends the classic method, resulting in a general inferential procedure that inherits most of the desirable features of traditional maximum likelihood, and at the same time gains some new properties that can be exploited in particular estimation settings.

        2 Generalized Nonlinear Model

        Suppose that the components ofY={y1,…,yn}Tare independent random variables, in which eachyimay depend on an independent known variablexi(i=1,2,…,n). The parameters of interest areβ={β1,…,βp}Tdefined in a subsetΒofRp(p

        g(μi)=f(xi,β),yi~ED(μi,σ2)i=1,2,…,n

        (7)

        whereg(·) is a known monotonic link function;f(·;·) is a known function with an unknown vector parameterβand a known explanatory variable ofq-vectorxi;E(yi)=μiand ED(μi,σ2) is the exponential family distribution, which has the density function in the following form as

        μi=μi(β) orθi=θi(β)

        (8)

        whereθiis the natural parameter; andσ2=φ-1is the dispersion parameter. According to the property of the exponential family distribution[7], we have

        (9)

        From model (8), the log-likelihood ofYfor the parameterβandφis usually denoted by

        (10)

        Then theLq-likelihood ofYis

        (11)

        (12)

        3 Diagnostics for Regression Parameter and Dis-persion Parameter

        yi~ED(μi,σ2)i=1,2,…,n

        g(μj)=f(xj;β)j=1,2,…,n;j≠i

        (13)

        This is just model (7) with thei-th case deleted.

        1) Generalized Cook distance

        (14)

        2) Likelihood distance

        The likelihood distance is defined as[1]

        (15)

        3) Difference of deviance

        The differences of deviance has the form as

        (16)

        whereD(β)dj(yj,μj(β)),D(i)(β)=dj(yj,μj(β)),dj(yj,μj(β))=-2{yjθj-b(θj)-c(yj)}+2{yjθj-b(θj)-c(yj)}μj=yj.ΔiDactually reflects the differences of the maximum likelihood estimators ofσ2orφunder the original model and under the CDM.

        4 Numerical Simulations

        We obtain the formula to the maximumLq-likelihood estimators ofβandφunder Eq.(12) in Section 2. In Section 3, three kinds of distance, which are used for the diagnostics, are introduced. Now let us look at a numerical simulation example for computing diagnostic statistics.

        Fig.1 GDi based on MLE. (a) n=30; (b) n=50; (c) n=80

        Fig.2 GDi based on the maximum Lq-likelihood estimator. (a) n=30; (b) n=50; (c) n=80

        Fig.3 LDi based on MLE.(a) n=30; (b) n=50; (c) n=80

        Fig.4 LDi based on the maximum Lq-likelihood estimator. (a) n=30; (b) n=50; (c) n=80

        From the above figures, we can obtain the same strong impact points through calculating the value about GDiand LDi. The value of GDiabout the strong impact point which obtains from the maximumLq-likelihood method is bigger than the value obtained from the maximum likelihood method when the sample size is small. With the increase in the sample size, the difference between them is small. It means that the maximumLq-likelihood estimation method is more effective than the MLE method when the sample size is small.

        5 Illustrative Example

        We have the formula to the maximumLq-likelihood estimators ofβandφin Section 3 and three diagnostic statistics in Section 4. Now let us look at an example for computing diagnostic statistics.

        Example1Product sales data

        These data were given by Whitmore[12], which is shown in Tab.1. We call them the “Produce sales data” for short. In this dataset,xirepresents the projected sales amounts of thei-th product reported by a market survey organization andyiis the corresponding actual sales amounts of a company (i=1,2,…,20). Whitmore suggested an inverse Gaussian fit by using

        (17)

        Then, we obtain the values of the diagnostic statistics (see Tab.2).

        All the results indicate that case 10 may be an outstanding outlier and the influence of case 20 is also obvious. Through the comparison between our results and the diagnostic results obtained through MLE[13], we find that No. 10 and 20 are two strong impact points, and the value of GDiand LDiusing the maximumLq-likelihood estimation method is greater than the value using the MLE method. It means that the result we arrived at is more significant. So it is feasible to diagnose using the maximumLq-likelihood method.

        Tab.1 Product sales data

        Tab.2 Some diagnostic statistics for product sales data

        6 Conclusion

        In this paper, we consider the diagnostics of generalized nonlinear regression models. Three diagnostic statistics and a new estimation method, the maximumLq-likelihood estimator, are introduced. Through the value of the diagnostic statistics, we can make a decision of whether the individual point of the data is a strong impact one or not. By comparing the results obtained from the maximum likelihood estimator and the maximumLq-likelihood estimator, we find that the method of the maximumLq-likelihood estimation is more effective when the sample size is small. When the sample size is bigger, the values about the diagnostic statistics are almost the same. In other words, the performance of our proposed method is equivalent to that of the classical diagnostic method for large sample sizes. The same conclusion can also be obtained from the example in Section 5.

        This paper discusses some diagnostic problems for the generalized nonlinear model by using the maximumLq-likelihood estimation method and compares the results with those of the classical diagnostic method. However, consistency and asymptotic normality of the maximumLq-likelihood estimation for the generalized nonlinear model have not been proved, which is obviously of great significance. Thus, more research on this issue will be valuable and we believe this is an interesting direction for further exploration.

        [1]Cook R D, Weisberg S.Residualandinfluenceinregression[M]. London: Chapman and Hall, 1982.

        [2]Chatterjee S, Hadi A S. Influential observations, high leverage points and outliers in linear regression (with discussion) [J].StatisticalScience, 1986,1(3): 379-416.

        [3]McCullagh P, Nelder J A.Generalizedlinearmodels[M]. London: Chapman and Hall, 1989.

        [4]Davison A C, Tsai C L. Regression model diagnostics [J].InternationalStatisticalReview, 1992,60(3): 337-355.

        [5]Gay D M, Welsch R E. Maximum likelihood and quasi-likelihood for nonlinear exponential family regression models [J].JournaloftheAmericanStatisticalAssociation, 1988,83(404): 990-998.

        [6]Wei B C and Shi J Q. On statistical models in regression diagnostics [J].AnnInstStatistMath,1994,46(2):267-278.

        [7]Wei B C.Exponentialfamilynonlinearmodels[M]. Singapore: Springer, 1998.

        [8]Ferrari D, Yang Y. Estimation of tail probability via the maximumLq-likelihood method [R]. Minneapolis, MN, USA: School of Statistics, University of Minnesota, 2007.

        [9]Kullback S.Informationtheoryandstatistics[M]. Wiley: New York, 1959.

        [10]Shannon C E. A mathematical theory of communication [J].BellSystemTechnicalJournal, 1948,27:379-423.

        [11]Havrda J, Charvt F. Quantification method of classification processes: concept of structural entropy [J].Kibernetika, 1967,3: 30-35.

        [12]Whitmore D A. Inverse Gaussian ratio estimation [J].JournaloftheRoyalStatisticalSociety,SeriesC(AppliedStatistics), 1986,35(1):8-15.

        [13]Wei B C, Lin J G, Xie F C.Statisticaldiagnosis[M].Beijing: Higher Education Press, 2009. (in Chinese).

        国产一精品一aⅴ一免费| 久久综合九色欧美综合狠狠| 国产精品国产亚洲精品看不卡| 未满十八勿入av网免费| 欧美天欧美天堂aⅴ在线| 欧美伊人亚洲伊人色综| 人妻少妇精品视频一区二区三区 | 人妻一区二区三区免费看| 国产一区二区在线免费视频观看| 丰满少妇被猛烈进入高清播放| 内地老熟女老少配视频| 美女扒开内裤让我捅的视频| 国产午夜av秒播在线观看| 男女爽爽无遮挡午夜视频| 亚洲AV伊人久久综合密臀性色| 久久免费网站91色网站| av在线免费观看男人天堂| 欧美亚洲一区二区三区| 国产综合无码一区二区色蜜蜜| 2021年最新久久久视精品爱| 国内人妖一区二区在线播放| 日本熟女精品一区二区三区| av色欲无码人妻中文字幕| 性大片免费视频观看| 玩两个丰满老熟女| 黄色大片一区二区中文字幕| 少妇精品偷拍高潮少妇在线观看| 亚洲中文无码av永久| 免费看黄a级毛片| 亚洲午夜无码AV不卡| 国产目拍亚洲精品二区| 午夜国产视频一区二区三区| 国产福利视频一区二区| 亚洲男人av香蕉爽爽爽爽| 97国产精品麻豆性色| 亚洲国产精品国自产拍性色 | 精品国产一区二区三区性色| 日本最新免费二区三区| 久久久久国产精品免费免费搜索| 亚洲av永久无码精品成人| 国产老熟女伦老熟妇露脸|