 Methodology
 Open Access
 Published:
On studentized residuals in the quantile regression framework
SpringerPlus volume 5, Article number: 1231 (2016)
Abstract
Although regression quantiles (RQs) are increasingly becoming popular, they are still playing a second fiddle role to the ordinary least squares estimator like their robust counterparts due to the perceived complexity of the robust statistical methodology. In order to make them attractive to statistical practitioners, an endeavor to studentize robust estimators has been undertaken by some researchers. This paper suggests two versions of RQs studentized residual statistics, namely, internally and externally studentized versions based on the elemental set method. The more preferred externally studentized version is compared to the one based on standardized median absolute deviation (MAD) of residuals using a wellknown data set in the literature. While the MAD based outlier diagnostic seemed to be uniform and more aggressive to flagging outliers the RQ externally studentized one exhibited a dynamic pattern consistent with RQ results.
Background
Tukey (1979) recommends that it is perfectly proper to routinely use both the ordinary least squares (OLS) and robust estimators and only examine the data more closely in case of “large” discrepancieswhatever this means (but it is widely accepted that this means that otherwise it suffices to use the OLS). However, this is rarely done as robust estimators are still playing a second fiddle role to the OLS estimator, despite their proliferation. The main reason why this status quo remains is that at the interface of statistics and its applications there are nonspecialists who find it insurmountable to deal with this vague idea of “large” discrepancies and the necessary choices of types of estimators and tuning constants involved in the robust statistical methodology. On the other hand the OLS has a clear and easy to implement methodology to conduct inference and goodness of fit analysis (including residual diagnostics). To make the robust estimators more appealing to statistical practitioners, an endeavor to studentize robust estimators has been undertaken by some researchers (see e.g. Mckean and Sheather 1991; Yohai et al. 1991). This studentization enables users to undertake pertinent statistical tests and obtain confidence intervals and critical values as well as outlier diagnosis which parallel the OLS ones.
Outliers (unusual observations in the Yspace) can adversely influence the regression model fit thereby invalidating the pertinent statistical inferences (see e.g. Rousseeuw and Leroy 2003; Barnett and Lewis 1998). The Koenker and Basset (1978) regression quantiles (RQs) are fairly robust to outliers as their influence functions are bounded in the Yspace. As a result, not only have RQs been employed as alternatives and complementary tools to the OLS estimator but also in robust outlier detection techniques (Portnoy 1991). These detection methods are based on a twofold approach, namely, the “peeling” of observations fit exactly by extreme RQs and those based on RQ computation, i.e., observations lying below the RQs hyperplanes \({{\widehat{q}}_{Y{\mathbf {x}}}}(\tau )\) and/or lying above \({{\widehat{q}}_{Y{\mathbf {x}}}}(1\tau )\) corresponding to \(\widehat{\varvec{\beta }}(\tau )\) and \(\widehat{\varvec{\beta }}(1\tau )\), \(\tau \in (0,1)\), respectively (see expression (7)) may be identified as outliers. Complemented by the ordinary least squares (OLS) one consequence of the latter approach is the Ruppert and Carroll (1980) regression trimmed mean estimator. Outliers in the Xspace are referred to as high leverage points. A worse outcome can result if outliers are further coupled with high leverage points in a data set than when either data aberration manifests alone, especially in the case of RQs. This stems from the fact that RQs are very susceptible to high leverage points since their influence functions are unbounded in the Xspace. This curtails their effectiveness to detect outliers that are also high leverage (outlierleverage) points due to the not yet so wellperceived tradeoff between the RQs high affinity for high leverage points and their exclusion of (resistance to) outliers. Studentization may be a solution as it involves incoorperating some Xinformation.
Most of the existing outlier diagnostics in the RQ framework are in relation to the global orientation (centre) of the data and not relative to each quantile level \(\tau \in (0,1)\), i.e., a conditional quantile model, \({{Q}_{YX}}(\tau )\), especially extreme ones. Very few quantile level specific diagnostics exist. One such single case outlier diagnostic in existence is based on the standardized median absolute deviation (MAD) of residuals (Huber and Ronchetti 2009). Given that it is wellknown that regression outlier diagnostics do not always agree in flagging outliers the conventionally agreed practice of employing a wide spectrum of diagnostics before the analyst arrives at a verdict cannot be exercised in the RQ framework. The focus of this paper is to contribute by adding some new outlier diagnostics to the few existing ones in the RQ framework and further bring in the OLS’s attractiveness to this framework via studentization of residual statistics. This is a convenient approach as RQs have a common link with the OLS estimator that can be fruitfully exploited. This link exists via the elemental set (ES) method (Hawkins et al. 1984). So a studentized residual statistics are suggested for RQs here based on the ES method.
An ES consists of exactly the minimum number (p) of observations to fit the regression model parameters. Such a proposal is motivated by the fact that the basic optimal solution of a linear programming (LP) problem giving a RQ coincides with the p points of an ES (see Koenker and Basset 1978, Theorem 3.1; Ranganai 2016). Applying the OLS procedure to the p ES observations yields a specific elemental regression (ER). Thus RQ leverage and residual statistics and ER ones are identical. A deterrent to employing the ES method is the possibly huge load involved in computing all the \(K=\left( {\begin{array}{c}n\\ p\end{array}}\right)\). However, the number of LP optimization solutions giving RQs is approximately equal to \(n<K\). Thus the ES approach benefits from the existence of efficient LP optimization algorithms giving RQs as solutions. Also, it is shown that the suggested RQ studentized residual statistics follow a t distribution from which a wide spectrum of cutoff values can be obtained like their OLS based counterparts. These are desirable attributes for the practitioner.
In summary the motivations for the development of studentized outlier diagnostics in RQ frame work, are the following:

Very few RQ \(\tau\) level specific outlier diagnostics with the efficacy to deal with all outlier configurations currently exist in the literature. Therefore the conventionally accepted practice of employing a wide spectrum of diagnostics cannot be carried out in the RQ framework unless more get developed.

Use of of efficient LP algorithms lessens the possibly huge load involved in computing all the K ESs as approximately \(n<K\) RQs from the LP solutions are of interest to this study.

Ease of implementation via OLS and the existence of a wide spectrum of cutoff values from the t distribution brings in the attractive of OLS to practitioners.

There is need to develop more single case outlier diagnostics in light of the not so well perceived opposing phenomena between outlier and high leverage behaviours in outlierleverage points.

Outlierleverage points may be identified better using outlier diagnostics as the suggested studentized diagnostics have some leverage (X information) inherent in them unlike the entirely residual (Y information) based ones.
Motivated by this background, this paper suggests outlier diagnostics based on studentization and ER. The rest of the paper is organized as follows; Some OLS leverage statistics and residuals are elaborated on in the next section; RQ leverage statistics and residuals are discussed in “Regression quantiles leverage statistics and residuals” section; “Studentized residuals in the quantile regression scenario” section dwells on the construction of the suggested RQ studentized residual statistics; Applications are given in “Applications” section while conclusions are given in the last section.
Some OLS leverage statistics and residuals
Consider the linear regression model,
where \({\mathbf{Y}}\) is an \(n\times 1\) vector of response observations, \({\mathbf{1}}_n\) is an \(n \times 1\) vector of ones, \({\mathbf{X}}\) is an \(n\times (p  1)\) matrix of predictor variables, \(\varvec{\beta }\) is a \((p  1) \times 1\) vector of regressors, \(\varvec{\varepsilon }\) is an \(n\times 1\) vector of errors, \(\varvec{\varepsilon } \thicksim N_n \left( {\mathbf{0}}_n, \sigma ^2 {\mathbf{I}}_n \right)\), \({\mathbf{0}}_n\) is an \(n \times 1\) vector of zeros, and \({\mathbf{I}}_n\) is an \(n \times n\) identity matrix. The \({{i}{\text {th}}}\) OLS residual is given by
where \(\widetilde{{\mathbf{x}}}_i' = \left[ {1}, {{\mathbf{x}}_i'} \right]\) with \({{\mathbf{x}}_i'}\) denoting the \({{i}{\text {th}}}\) row of \({\mathbf{X}}\). It is wellknown in the literature that the analysis of (raw) residuals (2) is far less potent in flagging outliers than the analysis of their transformed versions.
There are four versions of transformed residuals most frequently employed to identify outliers in the literature. We list them here in order of increasing effectiveness. These are the normalized, the standardized, the internally studentized and externally studentized residuals. The standardized OLS residuals are given by
where \(\widehat{\sigma }= \sqrt{MSE}\) with \(MSE=SSE/np\) and SSE denoting the usual OLS sum of squares of the error terms. Substituting \(\widehat{\sigma }\) in (3) by \(\sqrt{Var({{e}_{i}})}=\widehat{\sigma }\sqrt{ 1{{h}_{i}}}\) yields the internally studentized residuals
where \({{h}_{i}}\,=\,\,{{\widetilde{\mathbf {x}}}'_{i}}{{\left( {{{\widetilde{{\mathbf {X}}}'}}_{{}}}{{{\widetilde{\mathbf {X}}}}_{{}}} \right) }^{1}}{{\widetilde{\mathbf {x}}}_{i}}\), the \({{i}{\text {th}}}\) diagonal element of the hat matrix \(\mathbf {H}=\widetilde{\mathbf {X}}{{\left( \widetilde{\mathbf {X}}'\widetilde{\mathbf {X}} \right) }^{1}}{\widetilde{\mathbf {X}}'}\) denotes the leverage of the \({{i}{\text {th}}}\) observation. Under model (1) assumptions, \({{t}_{i}}\) follows a t distribution with \(np\) degrees of freedom, i.e., \({{t}_{i}} \sim {{t}_{np}}\).
Finally, the externally studentized residuals follow from substituting \(\widehat{\sigma }\) in (3) by \(\sqrt{Var({{e}_{(i)}})}=\widehat{\sigma }_{(i)}\sqrt{(1{{h}_{i}})}\), where the subscript notation (i) indicates the deletion of the \({{i}{\text {th}}}\) observation and \({\widehat{\sigma } }_{(i)}^{2}=[(np)\widehat{\sigma }^{2}{e_{i}^{2}}/{(1{{h}_{i}})}\;]/[{np1}],\) giving
Also, like \({{t}_{i}}\), under model (1) assumptions \({{t}_{(i)}}\) follows a t distribution with \(np1\) degrees of freedom, i.e., \({{t}_{(i)}} \sim {\ }t(np1)\)
Another version of the residuals that is often used to assess prediction are the jackknife (predicted) residuals
The jackknife residuals have been found to be more effective than the OLS ones in assessing prediction and flagging outliers in the literature (see e.g. Myers et al. 2010). The predicted sum of squares gives the wellknown PRESS statistic,
In the next section some of the analogues of the OLS statistics discussed here are adapted to the RQ scenario.
Regression quantiles leverage statistics and residuals
The \({{\tau }{th}}\) RQ based on the linear model is a solution to the linear programming (LP) problem
where \({{\rho }_{\tau }}(u)=u[\tau I(u<0)]\equiv u[\tau \cdot I(u\ge 0)+(\tau 1).I(u<0)]\), for \(\tau \in (0,1)\). The basic optimal solution to this LP problem (7) obtained using efficient LP algorithms in the literature, is a RQ that corresponds to a specific ES of size p (see Koenker and Basset 1978, Theorem 3.1, p. 39; Koenker 2005, Subsection 2.2.1). Two major linear programming techniques exist for solving the above linear programming problem, viz., exterior and interior methods.
Letting \(\widetilde{{\mathbf{X}}} = \left[ {\mathbf{1}}_n, {\mathbf{X}} \right]\) in terms of ESs the linear model (1) can be expressed as
where \(\widetilde{{\mathbf{X}}}_J\) is \(p\times p\) and \(\widetilde{{\mathbf{X}}}_I\) is \((np)\times p\) matrices. Let \(\left( \begin{array}{cc} \widetilde{{\mathbf{X}}}_J&{\mathbf{Y}}_J \end{array} \right)\) be a generic ES, then \(K = {n \atopwithdelims ()p}\) is the number of ESs. The subset J corresponds to the set of subscripts \(\{{{h}_{1}},...,{{h}_{p}}\}\) such that \(({{\mathbf {x}}'_{{{{{hi}}}}}},{{y}_{hi}})\), \(i=1,...,p\), is the the \({{i}{\text {th}}}\) case of ES J. Applying OLS to an ES based on a subset J of size p of the original data results in the following vector of regression coefficients estimates
where \(\widetilde{{\mathbf{X}}}_J\) is a square matrix and assumed to be nonsingular. Since a RQ solution of (7) corresponds to ER (8) then their leverage statistics and residuals are identical.
RQ/ER leverage statistics are the diagonal elements of the matrix \({{\mathbf {H}}_{J}}=\widetilde{\mathbf {X}} ({\widetilde{\mathbf {X}}}'_{J}{\widetilde{\mathbf {X}}}_{J})^{1} \widetilde{\mathbf {X}}'\), i.e.,
The statistic \(h_{iJ},\; i \not \in J\) is referred to as the ER predicted (ERP) leverage. Note that this statistic is the jackknife analogue of the \({{i}{\text {th}}}\) diagonal element \({{h}_{(i)}}\,=\,\,{{\widetilde{\mathbf {x}}}'_{i}}{{\left( {{{\widetilde{\mathbf {{X}}'}}}_{(i)}}{{{\widetilde{\mathbf {X}}}}_{(i)}} \right) }^{1}}{{\widetilde{\mathbf {x}}}_{i}},\) of another variant of the hat matrix \({{\mathbf {H}}_{(i)}}=\widetilde{\mathbf {X}}{{\left( {{{{\widetilde{\mathbf {X}}'}}}_{(i)}}{{{\widetilde{\mathbf {X}}}}_{(i)}} \right) }^{1}}{\widetilde{\mathbf {X}}'}\).
The RQ/ER residuals are given by
The residuals \(e_{iJ},\; i \not \in J\) which are the analogues of the jackknife (predicted) residuals (6) are referred to as elemental predicted residuals (EPRs). EPRs have has variance
Following from this variance, Hawkins et al. (1984) referred to \({{h}_{iJ\,\,}},\,\, i\notin J\) as the residual freedom, to “convey the impression of its property of measuring the extent to which the elemental set J fails to predict \({{Y}_{i}}\).” Consequently \({{{e}_{iJ}}}/\sigma {\sqrt{1+hiJ}}\;,\quad i\notin J\sim {\ }N(0,1)\).
Summing the EPRs gives the analogue of the PRESS statistic
Residual analysis in the ER case is redundant, since the ER (internal) residuals suffer from the exact fit property, i.e., the (internal) residuals are constants (zeros), and hence, the same applies for the RQ case. However, the external ones, i.e., ER predicted (ERP) residuals which are the analogues of the jackknife (leave one observation out) residuals are useful. Similarly ERP leverage is also useful. Thus in the next section RQs studentized residuals are constructed using ERP residuals and ERP leverage values.
Studentized residuals in the quantile regression scenario
In this section we construct a version of studentized residuals for RQs. We do this by first suggesting a scaled version of the RQ predictive residuals (EPRs),
where \(J_\tau\) denotes the ES corresponding to the \({{\tau }{th}}\) RQ for \(\tau \in (0,1)\) since we are only interested in RQs (ESs corresponding to RQs). The statistic \({{\widehat{\sigma }}_{(J_\tau )}}\) is the scaled prediction variance with the p observations left out corresponding to a RQ (ER) \(J_\tau\) left out, i.e.
where \(PRES{{{S}'}_{J_\tau }}=\sum \nolimits _{i\notin J_\tau }{e_{iJ_\tau }^{2}/(1+{{h}_{iJ_\tau }})\ }\) and \(\alpha =2p\) accounting for the p parameters as well as the p ER observations left out corresponding to \({e}_{iJ_\tau }=0\) for \(i\in J_\tau\). In line with the literature convention the RQ externally studentized residuals or externally studentized EPRs (SEPRs) should be based the jackknife residual variance
i.e., with the \({{i}{\text {th}}}\) observation left out. This statistic is given by
where \({{\widehat{\varepsilon }}_{iJ_\tau }}={{{e}_{iJ_\tau }}}/{\sqrt{1+hiJ_\tau }}\;,\quad i\notin J_\tau\) to flag outliers. The internally studentized version is given by
The distributions of the these statistics ((14) and (15)) are given by Theorems 1 and 2 from which we determine the appropriate cutoff values.
Theorem 1
Under model (1) the RQ externally studentized residuals \({{\upsilon }_{(i)J_\tau }}\sim {\ }t(n2p1)\).
Proof
Let \({{\theta }_{i}}={{t}_{iJ_\tau }}\sqrt{1+hiJ_\tau },\quad i\notin J_\tau ,\) with \({{t}_{iJ_\tau }}=\frac{{{e}_{iJ_\tau }}}{{{{\widehat{\sigma }}}_{(J_\tau )}}\sqrt{1+{{h}_{iJ_\tau }}}}=\frac{{{{\widehat{\varepsilon }}}_{iJ_\tau }}}{{{{\widehat{\sigma }}}_{(J_\tau )}}}.\)
Substituting (12) into \({{\theta }_{i}}\), we have
Therefore
with \({{\widehat{\varepsilon }}_{iJ_\tau }}\). So \(0\le \frac{\widehat{\varepsilon }_{iJ_\tau }^{2}}{PRES{{{{S}'}}_{J_\tau }}}\le 1\) and \(\theta _{i}^{2}\le (n\alpha )\) or equivalently \({{\theta }_{i}}\le \sqrt{(n\alpha )}\) meaning that the density function of \({{\theta }_{i}}\) is zero outside \([\sqrt{(n\alpha )},\sqrt{(n\alpha )}].\) Now let
The second factor can be simplified as
The denominator component in the square root sign can be expressed as
where \(PRES{{{S}'}_{(i)J_\tau }}=\sum \nolimits _{j\ne i}{e_{jJ_\tau }^{2}/(1+{{h}_{jJ_\tau }})\ =\sum \nolimits _{j\ne i}{\widehat{\varepsilon }_{jJ_\tau }^{2}\ }},\,\,\text {for}\quad i,j\notin J_\tau .\) Then
Multiplying this result by the first factor in \({{\upsilon }_{iJ_\tau }}\) we have
Therefore
since \(\frac{{{{\widehat{\varepsilon }}}_{iJ_\tau }}}{{{\sigma }}}={{{e}_{iJ_\tau }}}/\sigma {\sqrt{1+hiJ_\tau }}\sim {\ }N(0,1)\) and \({\frac{1}{\sigma ^{2}(n\alpha 1)}PRES{{{{S}'}}_{(i)J_\tau }}}\sim {\ }{{\chi }^{2}}(n\alpha 1).\) Taking \(\alpha =2p\) we have
\(\square\)
Theorem 2
Under model (1) the RQ studentized internally residuals \({{\upsilon }_{iJ_\tau }}\sim {\ }t(n2p)\).
Proof
The proof follows from that of Theorem 1 by substituting \({(n\alpha 1)^{1}}PRES{{{{S}'}}_{(i)J_\tau }}\) with \({(n\alpha )^{1}}PRES{{{{S}'}}_{J_\tau }}\) for the estimated EPR variance. Thus the final result becomes
since \(\frac{{{{\widehat{\varepsilon }}}_{iJ_\tau }}}{{{\sigma }}}={{{e}_{iJ_\tau }}}/\sigma {\sqrt{1+hiJ_\tau }}\sim {\ }N(0,1)\) and \({\frac{1}{\sigma ^{2}(n\alpha )}PRES{{{{S}'}}_{J_\tau }}}\sim {\ }{{\chi }^{2}}(n\alpha ).\) Taking \(\alpha =2p\) we have
\(\square\)
Therefore the appropriate Bonferroni critical values are \(t(1\alpha /2(np);n2p1).\) The advantage of these critical values is that the Bonferroni method is simple and allows many comparisons to be made simultaneously while still maintaining an overall confidence coefficient. In the literature externally studentized diagnostics are shown to outperform their internal versions counterparts. Therefore it is preferred here to compare the externally SEPR \({{\upsilon }_{(i)J_\tau }}\)’s outlier flagging pattern to the MAD version in the SAS QUANTREG procedure. Using the MAD based version of the RQ predicted residuals, outliers are identified as
where the multiplier k usually takes values, 3, 4 or 5. The scale parameter \({{\widehat{\sigma }}_{m}}\) is the corrected median of absolute values \({{\widehat{\sigma }}_{m}}=\text {median}\left\{ {{e}_{iJ_\tau }}/{{\theta }_{0}},1\le i\le n \right\}\), where \({{\theta }_{0}}={{\Phi }^{1}}(0.75)\) is an adjustment consistency with the normal distribution.
In the next sections the flagging rate of outliers based on this cutoff value in expression (16) and the ones from (14) based on critical values of the t distribution are compared using the Hocking and Pendleton (1983) data set.
Applications
In this Section we consider the Hocking and Pendleton (1983) data set. This data set is a plausible candidate to study the efficacy of the SEPR in flagging outliers as it has various various outlier and high leverage scenarios that are both easy and challenging to deal with in the RQ framework. These include a very high leverage observation 24, an outlier in 17 and two outlierleverage points 11 and 18 with varying degrees of high leverage. Observation 24 will almost always be included in the ES corresponding to RQs due to RQs affinity for high leverage points. Thus it will often have a zero residual while observation 17 will almost always be excluded in this ES and will often have a very large residual. The challenge is on outlierleverage points 11 and 18 which will depend on the tradeoff of the two antagonistic phenomena, namely, the RQs’ affinity for leverage points versus their exclusion (resistance) to outliers.
It is wellknown that externally studentized residual statistics always perform better than their internally studentized counterparts since (5) and (14) are based on \(\widehat{\sigma }_{(i)}\) and \(\widehat{\sigma }_{(i)({{J}_{\tau }})}^{2}\) which are both more robust to problems of gross errors in the \({{i}{th}}\) observation than \({{\widehat{\sigma }}^{2}}\) and \(\widehat{\sigma }_{(J_\tau )}^{2}\) on which (4) and (15) are based, respectively (Chatterjee and Hadi 1988, pg 79). Therefore the externally studentized residual criterion (14) is compared to the robust version one based the standardized MAD of residuals (16). Criterion (16) is the only single case similar RQ level related outlier diagnostic with which to validate the efficacy of (14). Firstly the robust and multivariate location and scale diagnostics computed using the minimum covariance determinant (MCD) method of Rousseeuw and Driessen (1999) are applied to circumvent the masking and swamping phenomena so as to expose all the single case high leverage points and outliers. The resulting diagnostic outcome is given in Fig. 1.
The flagging pattern based criteria MAD (16) and SEPR (14) for the Hocking and Pendleton data set are given in Table 1. For criterion (16) the multiplier k values were chosen as 3 (*) and 4 (**) while for criterion (14) the significance level \(\alpha ={0.10}\) was chosen so as to be both liberal and stringent in flagging outliers. The liberal and stringent Bonferroni cutoff values correspond to \({{\upsilon }_{(i)J_\tau }}>t(1{\alpha }/{2;n2p1)}=\pm {1.740}\) and \({{\upsilon }_{(i)J_\tau }}>t(1{\alpha }/{2(np);n2p1)}\pm {3.544}\), respectively.
Remark
ESs Corresponding to RQs are the \(p=4\) observations (with zero residuals) in the basic optimal solution of LP problem (7) obtained using effeicient linear programing algorithms.
The two outlier diagnostics do not always agree as is the norm in any regression diagnosis outcome using different diagnostics. Observation 24 with the highest leverage and non outlying is never flagged at all. The major difference to note here is the uniform flagging exhibited by (16) from \(\tau ={0.2046}\) to \(\tau ={0.8276}\) and only otherwise in very extreme \(\tau\) levels. It is hard to conceive that results for below and above \(\tau ={0.50}\) are similar to this extent. This is inconsistent with the wellknown outcome of RQ results due their ability to capture the changing conditional distribution of the response variable, Y given the predictor factors, X at different quantile levels (Chamberlain 1994; Cade and Noon 2003). On the other hand criterion (14) has a dynamic pattern consistent with RQs results as expected.
Conclusion
The version of the studentized RQ predicted residuals (SEPRs) suggested here are useful and of benefit to statistical practitioners as they add to the few existing single case outlier diagnostics in the RQ scenario. Further, the methodology is easy to implement as they have cutoff values that parallel the OLS based versions. Thus they offer alternatives to nonspecialists who may fight it too hard to comprehend the robust outlier detection methodology. However, if possible these diagnostics must be used together as recommended by Tukey (1979).
References
Barnett V, Lewis T (1998) Outliers in statistical data. Wiley, New York
Cade BS, Noon BR (2003) A gentle introduction to quantile regression for ecologists. Front Ecol Environ 1(8):412–420
Chamberlain G (1994) Quantile regression, censoring, and the structure of wages. In: Sims CA (ed) Advances in econometrics: sixth world congress, vol 1. Cambridge University Press, Cambridge, UK, pp 171–209
Chatterjee S, Hadi AS (1988) Sensitivity analysis in linear regression. Wiley, New York
Hawkins DM, Bradu D, Kass GV (1984) Location of several outliers in multipleregression data using elemental sets. Technometrics 26(3):197–208
Hocking RR, Pendleton OJ (1983) The regression dilemma. Commun Stat Theory Methods 12(3):497–527
Huber S, Ronchetti EM (2009) Robust statistics, 2nd edn. Wiley, New Jersey
Koenker R (2005) Quantile regression: econometric society monographs. Cambridge University Press, New York
Koenker R, Basset G (1978) Regression quantiles. Econometrica 46(1):33–50
Mckean JW, Sheather SJ (1991) Small sample properties of robust analyses of linear models based on Restimates: a survey. In: Stahel W, Weisberg S (eds) Directions in robust statistics and diagnostics Part II. The IMA volumes in mathematics and its applications, vol 34. Springer, New York, pp 1–19
Myers RH, Montgomery DC, Vining GG, Robinson TJ (2010) Generalized linear models: with applications in engineering and the sciences. Wiley, New Jersey
Portnoy S (1991) Regression quantile diagnostics for multiple outliers. In: Stahel W, Weisberg S (eds) Directions in robust statistics and diagnostics Part II. The IMA volumes in mathematics and its applications, vol 34. Springer, New York, pp 145–157
Ranganai E (2016) Quality of fit measurement in regression quantiles: an elemental set method approach. Stat Probab Lett 111(2016):18–25
Rousseeuw PJ, Leroy AM (2003) Robust regression and outlier detection, 3rd edn. Wiley, New Jersey
Rousseeuw PJ, Van Driessen K (1999) A fast algorithm for the minimum covariance determinant estimator. Technometrics 41(3):212–223
Ruppert D, Carroll RJ (1980) Trimmed least squares estimation in the linear model. J Am Stat Assoc 75(372):828–838
Yohai VJ, Stahel W, Zamar RH (1991) A procedure for robust estimation and inference in linear regression. In: Stahel W, Weisberg S (eds) Directions in robust statistics and diagnostics Part II. The IMA volumes in mathematics and its applications, vol 34. Springer, New York, pp 365–374
Acknowlegements
The author appreciates the Editor, the Associate Editor and the reviewers inputs which greatly improved the paper as well as the University of South Africa for funding this research.
Competing interests
The author declares that he has no competing interests.
Funding
The research was supported by the University of South Africa’s Research Department.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Ranganai, E. On studentized residuals in the quantile regression framework. SpringerPlus 5, 1231 (2016). https://doi.org/10.1186/s4006401628986
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s4006401628986
Keywords
 Leverage
 Outlier
 Studentized residual
 Regression quantiles
 Elemental set
 Elemental regression
 Elemental predictive residual