Ialongo: Confidence interval for quantiles and percentiles


Percentiles and quantiles are statistics used for summarizing the relative location of data within a set according to their magnitude and independently from a particular underlying probability distribution (1). Owing to this, they are broadly applied in biomedical field where non-normality due to outlier contamination or natural skewness is commonly observed.

Percentiles are also useful tools in the field of quality management to show the distribution of observed performance data and for attributing quality grading and goals in extra-analytical processes through indicators (2). A set of central percentiles that partition the population into equally sized ranges of values (e.g. the 25th, 50th and 75th percentiles collectively known as “quartiles”) are commonly employed to attribute a progressively higher level of performance (3). Another quality application concerns establishing reference intervals for interpreting laboratory tests results (4). In this scenario, a pair of extreme percentiles, often 0.5th - 99.95th or 2.5th - 97.5th that cut-off 1% or 5% population, respectively, are used to find those subjects whose testing seems to exceed the expected homeostatic variability in a biological parameter (5).

In both of these applications, a sample is drawn once from the population to find out the estimate of the true parameter, afterwards the value is repeatedly applied to a number of new individuals or items to identify them. According to theory, any point estimate is bound to its sample by an instant bias that depends on the randomness that occurred at the time of the sampling process (6). Therefore, any new observation randomly withdrawn from the same population is not necessarily compatible with the former point estimate because bias was not necessarily the same. Of course, the issue reduces to show that the newly observed value did not differ significantly from the sample estimate and in turn the true population parameter. To this end, it is suitable using the frequentist confidence interval (CI) whereby it is found the range of putative population true value that did not differ from the sample estimate with a confidence level as large as 1-α (7, 8).

To date, the discussion about using CI on extreme percentiles in building reference intervals has spurred investigations and recommendations that have been included in guidelines issued by official bodies and reviewed in books (9-17). However, little has been done concerning central percentiles, which are pivotal for the undergoing development of the extra-analytical quality field. Therefore, the present work was meant to give a theoretical introduction to the topic, also providing a comparison of the methods suitable for placing CI on percentiles via parametric, non-parametric and resampling approaches. To this end, we made use of numeric simulations for reproducing various conditions commonly encountered in laboratory medicine quality management where the departure from normality is more or less pronounced but the true population distribution is seldom known.

A technical premise on percentiles and quantiles

First of all, let’s start recalling that the quantile (xr) is the rth element within an ordered set of size N whose value is larger than or equal to that of q = r/N elements (i.e. x1 ≤ x2 ≤…≤ xr…≤ xn-1 ≤ xn). According to the frequentist paradigm, the probability (P) that any observation xi within the set has to occur can be defined with respect to xr with the following equation (Eq.):

P(xi ≤ xr) = q (Eq. 1).

If certainty is expressed as 100% of occurrence of observations, then it can be written that p = (100*q)% is the percentile of the dataset (1). Since the distinction between percentiles and quantiles reduces to the indexing, then methods discussed in the next sections are equally valid for both of them, even though they are presented using the quantile and thus xr.

The parametric CI method (in the Gaussian case)

Several parametric approaches can be used in order to estimate CI about the sample quantile (P-CI) when the underlying distribution is of the Gaussian kind (18). To this concern, if the dataset had average m and standard deviation s, the value xr could be sought straightforwardly via the standardization procedure:

z = (xr - m) / s (Eq. 2).

In fact, rearranging Eq. 2 yields:

xr = m + (z*s) (Eq. 3).

Notably, since n is a sample from the population N, m and s were estimates of the true parameters µ and σ respectively. Accordingly, xr was the estimate (^xr) of the population true quantile Xr giving the partition R = q/N. Therefore, we can write:

^xr = m + (z*s) (Eq. 4)
Xr = µ + (z*σ) (Eq. 5).

Particularly, Eq. 5 shows that wherever µ and σ were known also the true quantile Xr was so.

At this point it is possible to reason concerning the CI on ^xr (19). Of course, the estimate ^xr depended on both the sampling error (s) and the true value of the quantile Xr. If the latter was postulated basing on the assumption of normality and thus was given through the z of the standard normal curve N(0,1) according to Eq. 5 for the given percentile q, then we could summarize the “accuracy” of our finding through the quantity:

V = (^xr - Xr) / s (Eq. 6).

There is a striking similarity between Eq. 6 and the Student t-statistic, and indeed the V-statistic shows how ^xr varies around Xr that in turn is how the estimate difference (^xr - Xr) is distributing. To this regard, the V-statistic was shown to follow a Student t-distribution with n-1 degrees of freedom but with non-null centrality parameter (19). This particular condition (termed “non-centrality”) is explained by the evidence that the V-statistic was taken under the alternative hypothesis of non-null difference of the estimate from the true parameter since ^xr was assumed to be biased (i.e. deviating) by definition according to Eq. 4 (in contrast, the t-statistic is usually evaluated under the hypothesis of null difference and thus the non-centrality parameter equates 0) (18). Accordingly, the distribution of the V-statistic can be used to find out the range that ^xr was expected to lay in with probability of untrue finding equal to α:

P(a ≤ V ≤ b) = 1 – α (Eq. 7).

Therefore, rewriting Eq. 6 in the appropriate manner and substituting for it in Eq. 7, given the non-centrality parameter of the t distribution λ = - z*n0.5 (that we will refer to as t[n-1,λ]) it can be shown after simplifications:

lower P-CI = [m – (t1-α/2,[n-1,λ]*s*n -0.5)] (Eq. 8)
upper P-CI = [m – (tα/2,[n-1,λ]*s*n -0.5)] (Eq. 9).

Notably, whenever the underlying distribution is known, although not Gaussian (e.g. Weibull or lognormal), an appropriate parametric procedure to form CI can be set up likewise (20, 21).

The non-parametric CI method

Let us begin recalling that for q = r/n, then r = n*q (e.g. if q = 0.2 and n = 10 then r = 2) that is simply the number of elements within the subset identified by xr.

Supposing that n was withdrawn from a population N, obviously the sample quantity xr depended upon how many observations smaller than or equal to the population Xr were actually withdrawn. To this concern, it must me noted that each sample observation xi had chance q to be xi ≤ Xr (i.e. belonging to the partition of N as large as q*N), and thus it was known a-priori regardless of the true value Xr (22, 23).

According to the reasoning so far, the chance to find a certain estimate of Xr (say ^xr) can be reduced to the probability that sampling n consecutive and independent xi produced r findings smaller than or equal to Xr when this particular event had an individual chance as large as q. Noteworthy, this is nothing but the probability associated with n binomial trials (i.e. success in withdrawing xi ≤ Xr) and it can be easily sought whereby the binomial distribution with parameters n and q that is Bin(n,q) (23).

The first remark is that given n trials and the a-priori probability q of success, the expected outcome with the greatest chance to happen is exactly n*q = r. Therefore, r is the average of the sampling process which produced the dataset n and in turn the estimate ^xr. Secondly, random outcomes with
k ≠ r, and thus with larger or smaller number of observations xi ≤ Xr follow the distribution Bin(n,q).

The latter notion is noteworthy since it is useful for calculating the CI to be placed on xr. In fact, let us recall that by definition the CI is the interval that covers the putative true value of a population parameter with confidence 1-α, given the observed sample values. Therefore, given sample size of n and given the a-priori probability q, the CI can be stated in terms of realizations of binomial trials for kL ≤ r and kU ≥ r as follows:

P(kL ≤ r ≤ kU) ≈ 1-α (Eq. 10).

Hence, if xL and xU are the observations that cut off as many observations as kL and kU respectively, it can be written that:

P(xL ≤ xr ≤ xU) ≈ 1-α (Eq. 11).

Thus recalling Eq. 1 it yields (23):

P(xr ≤ xU) - P(xr ≤ xL) = Bin(xU;n,q) – Bin(xL;n,q) ≈ 1-α (Eq. 12).

Where Bin(xk;n,q) is the cumulative binomial distribution that enables the estimation of the probability of having up to k success in n trials given the a-priori probability q.

Several remarks are in order. First, Eq. 12 does not provide an exact coverage as it was for Eq. 7 since Bin(n,q) is a discrete distribution with only n+1 outcomes. More precisely, it tells that each value in the pair (xL;xU) must correspond to a fraction in the cumulative probability so that their difference is as close as possible to 1-α (and sometimes the symbol “≥” is preferred to “≈”). Therefore, (xL;xU) must be sought through an iterative procedure that attempts several alternative pairs and compares. Second, Eq. 12 only depends on the realization of q probability in n trials and not on the sample statistics. Therefore, this method does not require any assumption regarding the underlying distribution of data, so that it is regarded as distribution-free or non-parametric (NP-CI).

If sample size is adequate (usually n ≥ 20) it is possible to exploit the so-called normal approximation of the binomial distribution to simplify the procedure (24). In fact, it can be stated that:

lower NP-CI = (n*q) – zα/2*
((n*q)*(1 - q))0.5 (Eq. 13)
upper NP-CI = (n*q) + zα/2 *
((n*q)*(1 - q))0.5 (Eq. 14).

Where the term (n*q)*(1-q) in both equations is the standard deviation of the approximated normal distribution and zα/2 is the standardized value of the normal quantile (Eq. 2) cutting off the values in either tail whose cumulative probability is less than α/2. The equations above return the size of the subset that they respectively bound so that the rounded value corresponds to the position of the bounding value within the ordered set. For instance, if Eq. 13 yielded “6” then the lower NP-CI would correspond to the observation upper-bounding the subset of the least 6 observations. See Appendix for further details on calculation.

The resampling-based non-parametric (bootstrap) CI method

A third way for estimating CI about the quantile estimate involves data resampling (25). The underlying principle is quite obvious: if the random sample was produced by virtue of the “tendency” to follow a certain distribution in the original population, then the random sampling of the sample itself (i.e. the re-sampling) would reasonably replicate the very same tendency (26, 27).

Let us accordingly suppose that the “tendency” for a quantile q was represented by r = n*q, so that in absence of any sampling bias xi ≤ xr was actually xi ≤ Xr and the population estimate ^xr would be unbiased. In a real-world sample, randomness bias adds to the tendency in the population so that how many observations that were truly xi ≤ Xr remains unknown and ^xr is only a “particular” realization of Xr. Nonetheless, resampling the sample adds further bias to the former sampling process, so that the “second-generation” estimate ^xr would be differently biased.

If what is stated so far holds true for a single resample, actually randomness in resampling means that resampling bias may even be negative with respect to the original sampling bias. Therefore, if the resampling process was repeated a very large number of times (i.e. ≥ 1,000), then a distribution of bias about ^xr would arise (27). Therefore, the resampling distribution would show different proportions of observations xi ≤ xr, and thus of xi ≤ Xr, thereby allowing to gain information on the putative values for the true population parameter Xr. Notably, this is somehow similar to what seen previously for the NP-CI, where the binomial distribution was used to show alternative partitioning for the same sampling size n and the a-priori probability q.

The way in which it is used the information about the shape of the resampling distribution of Xr grounds the so-called non-parametric Bias Corrected-accelerated bootstrap method for building CI (BCa-CI) (28). Briefly, if α/2 and 1-α/2 are the percentiles of the resampling distribution providing 1-α coverage, indeed their centring and thus accuracy about the true population parameter is correct only when the resampling distribution is unbiased. Otherwise, it is necessary to adjust their “position” shifting the percentile boundaries from the original points to new ones that are the following:

lower BCa-CI = Φ (^z0 +
((^z0 + zα) / (1 - ^a*(^z0 + zα))) (Eq. 15)
upper BCa-CI = Φ (^z0 +
((^z0 + z1-α) / (1 - ^a*(^z0 + z1-α))) (Eq. 16).

where Φ denotes the cumulative standard normal distribution, zα and z1-α are the quantiles of the standard normal distribution, ^z0 and ^a are the parameters accounting respectively for resampling bias (i.e. the standardized proportion of ^xr > xr) and skewness (i.e. change of variance across the distribution from tail to tail when they are not symmetric about ^xr) (25, 28). Of course, the expected coverage probability remains 1-α. More comprehensive explanation on bootstrap resampling methods and related procedures to build CI can be found elsewhere (29-32).

Simulation study

The basic features concerning CI are represented by: a) the actual coverage probability or interval accuracy, b) the interval width and c) the interval centring or asymmetry (i.e. the ratio between the distances of the estimate from each of the boundaries). For the sake of conciseness we concerned mainly with point “a” as it is the one that should primarily be considered when making comparisons through different methods. Nonetheless, points “b” and “c” were considered only when the actual coverage probability was close to the nominal goal of 1-α.

To this end, we proceeded as follows: a) a theoretical model represented by the generalized 3-parameter lognormal distribution was used to generate sets of artificial data each featured by a combination of location (α = 0.5, 1.0, 2.0 and 3.0) scale (β = 0.05, 0.2, 0.5 and 1.2) and threshold (γ = 0) in order to reproduce a particular degree of asymmetry and tailing (i.e. skewness) with only positive values (X ≥ 0); b) for each set it was generated 3 batches of 100 samples sized n = 20, n = 60 and n = 120 respectively; c) for each combination of parameters the 2.5th, 25th, 50th, 75th and 97.5th percentiles were computed whereby the corresponding theoretical lognormal function obtaining the “true” or population value and the actual coverage probability was the proportion of sample CI that contained it. Particularly, CI were computed whereby Eq. 8 and Eq. 9 for P-CI, Eq. 13 and Eq. 14 for NP-CI, and Eq. 15 and Eq. 16 for BCa-CI. In order to evaluate and make appropriate comparisons on performance, the advisable optimum was represented by covering at least the 1-α nominal value which was set equal to 95% in this study.

All the calculations were performed using Excel 2010 (Microsoft Corp., Redmond, USA), except for BCa that was performed using SPSS 20.0 (IBM Corp., Armonk, USA) and data generation that was carried out exploiting the pseudo-random number generator embedded in Minitab 17 (Minitab Inc., State College, USA).

An electronic spreadsheet based on Microsoft Excel framework is provided in Supplementary material in order to allow automatic calculations of P-CI and NP-CI for up to 500 sample data, plus an additional file with Worked examples showing practical applications in "real-world" scenarios of the methods here described.


Data modelling

The combinations of scale and location parameters gave rise to the following data models (Figure 1): S1) for β = 0.05 the shape was almost Gaussian and thereby equivalent to a normal distribution with coefficient of variation CV ≈ 5% and shape changing from leptokurtic (i.e. peaky) by α = 0.5 to platykurtic (i.e. flat) by α = 3.0; S2) for β = 0.2 the shape was almost Gaussian and thereby equivalent to a normal distribution with coefficient of variation CV ≈ 20% and same changes in shape seen for S1 at the variation of α; S3) for β = 0.5 the shape was right-skewed and changed from minimal right-tailed and platykurtic by α = 0.5 to heavily right-tailed and platykurtic by α = 3.0; S4) for β = 1.2 the shape was left-fronted (i.e. almost no left tail) and changed from leptokurtic with short right-tailing by α = 0.5 to platykurtic long heavy right-tailing by α = 3.0.

Figure 1

Theoretical frequency distribution of the simulated data generated by the 3-parameter lognormal probability density function. By varying scale (β) and location (α) parameters with threshold (γ) fixed at 0 it is possible to reproduce different combinations of asymmetry, tailing and kurtosis (flatness) that give rise to the testing conditions described in the result section as S1 - S4. In general, increasing α gives more flatness to the shape while β more asymmetry, whereas γ relatively affects the degree of left-fronting since it constrains the distribution of data to a certain lower bound.


CI accuracy

Analysis showed four different scenarios with respect to the actual coverage probability. When shape was the kind of S1 (Figure 2), the P-CI performed better than both NP-CI and BCa-CI for n = 20 as well as n = 60, while the three were almost equivalent for n = 120. However, the P-CI was the only one to provide adequate coverage probability for both the extreme (2.5th and 97.5th) and the central percentiles (25th, 50th and 75th).

Figure 2

Actual coverage probability in the quasi-Gaussian model of data with low CV% (S1). The coordinates in brackets (β;α;n) represent scale (β) and location (α) parameters of the lognormal distribution (with 0 threshold) generating the artificial data for a given sample size (n). The solid black line shows the desired coverage probability (95% in this case) that is achieved when the symbol representing the parametric-CI (circle), nonparametric-CI (square) or BCa-CI (triangle) lays on it.


When the shape was the kind of S2 (Figure 3), the three CI-building procedures performed almost equally adequate for the central percentiles regardless of n, while only the BCa-CI was close to the goal of 1-α for the extreme percentiles when sample size was large (i.e. n = 120).

Figure 3

Actual coverage probability in the quasi-Gaussian model of data with high CV% (S2). The coordinates in brackets (β;α;n) represent scale (β) and location (α) parameters of the lognormal distribution (with 0 threshold) generating the artificial data for a given sample size (n). The solid black line shows the desired coverage probability (95% in this case) that is achieved when the symbol representing the parametric-CI (circle), nonparametric-CI (square) or BCa-CI (triangle) lays on it.


When the shape was the kind of S3 (Figure 4), the P-CI resulted in unsatisfactory performance regardless of n for any percentile except for the 25th. On the contrary, both NP-CI and BCa-CI performed satisfactorily for the central percentiles, but their coverage of the extreme ones was always slightly below the goal of 1-α even for n = 120.

Figure 4

Actual coverage probability in the right-skewed model of data with right tailing (S3). The coordinates in brackets (β;α;n) represent scale (β) and location (α) parameters of the lognormal distribution (with 0 threshold) generating the artificial data for a given sample size (n). The solid black line shows the desired coverage probability (95% in this case) that is achieved when the symbol representing the parametric-CI (circle), nonparametric-CI (square) or BCa-CI (triangle) lays on it.


Finally, when the shape was the kind of S4 (Figure 5), the P-CI always returned unreliable boundaries (e.g. negative values) or poor coverage depending on n. Also in this case, both NP-CI and BCa-CI always provided satisfactory performance for the 50th percentile regardless of the data shape while the capability to cover all three central percentiles was gained just for n = 120 and platykurtic distribution. Notably, extreme percentiles were never adequately covered regardless of shape and sample size.

Figure 5

Actual coverage probability in left-fronted model of data with left fronting (S4). The coordinates in brackets (β;α;) represent scale (β) and location (α) parameters of the lognormal distribution (with 0 threshold) generating the artificial data for a given sample size (n). The solid black line shows the desired coverage probability (95% in this case) that is achieved when the symbol representing the parametric-CI (circle), nonparametric-CI (square) or BCa-CI (triangle) lays on it.


CI width and shape

Whit respect to extreme percentiles, width and shape of P-CI was incomparable with NP-CI and BCa-CI because of poor accuracy in the latter two methods.

With respect to central percentiles, for samples generated according to models S1 and S2, all three building methods returned fairly symmetric intervals, which were sensibly narrower for the P-CI. However, when data where generated according to model S3 and S4, data skewness reflected into proportionally increasing width and symmetry by 25th to 75th percentile for both NP-CI and BCa-CI.

It is remarkable that there was negligible difference in the boundaries returned by NP-CI and BCa-CI for a given percentile, shape and size of the sample up to n = 120. That similitude was preserved up to the size n = 320 (data not shown), even though BCa-CI resulted slightly narrower despite being equally centred to NP-CI (Figure 6).

Figure 6

Example of actual coverage probability (CI accuracy) pattern by numeric simulation. It is shown an example of the 25th percentile CI for sample size n = 60 and parameters of the lognormal distribution scale β = 0.2, location α = 0.5, threshold γ = 0, computed according to the parametric (a), nonparametric (b) and BCa (c) method. Dotted grey line is the true population 25th percentile while solid black lines are the CI that do not contain it.



In this work, we described three methods for building CI on quantiles and percentiles and investigated their performance regarding different data shape and sample size. Foremost, we observed that NP-CI and BCa-CI almost equally behaved in all the situations herein explored, while P-CI was evidently something different because explicitly bounded to the normality of data. However, the major difference was due to the way the three procedures produced the CI boundaries. Indeed, the P-CI relied upon the distribution of the sample statistic and thus an external set of values derived by a theoretical distribution (the non-central t). By contrast, NP-CI as well as BCa-CI used observed values, and precisely the NP-CI did it directly by picking up value from the sample according to an alternative partitioning while the BCa-CI did it indirectly by recreating the alternative partitioning through resampling and picking up the values afterwards. Thereby, NP-CI and BCa-CI were constrained to return boundaries that were always within the range of the sample values.

Such technical aspects explain our major findings. Indeed, NP-CI and BCa-CI never returned CI boundaries with negative sign, which is crucial since we supposed to deal only with positive quantities. However, the analysis of the extreme percentiles demonstrated that their performance was dramatically affected by the sample size for they required n = 120 or even larger depending on the degree of skewness and tailing to achieve acceptable accuracy (i.e. ≈ 90%) (33, 34). In contrast, P-CI were accurate yet by n = 20 where the shape was close to normality (e.g. S1), showing the better accuracy of parametric approach (18, 22, 33). Notably, these findings seem to be contrasting with the recommendation of the International federation of clinical chemistry (IFCC) according to which CI for RI should be computed by means of bootstrapping (14). In our opinion, this can be explained considering that IFCC aimed at preventing inappropriate use of parametric techniques by favouring robustness of the procedure over accuracy of the method. In fact, in order to get reliable P-CI it is necessary to satisfy three conditions: 1) that the underlying distribution of data is correctly ascertained, 2) that the fitly P-CI procedure is available, and 3) that the parametric method is appropriately chosen for that distribution. The reader can refer to the worked examples provided in Supplementary materials for a more practical insight also preliminary data analysis. In this regard, we feel to suggest the Anderson-Darling as it is a goodness-of-fit test that quantifies the degree of deviation from normality whereby the AD statistics. Nonetheless, we also recommend using a normality plot to visually inspect potential local deviations in the body or tail of the data distribution, and thus Percentile-Percentile plot or Quantile-Quantile plot for central or extreme percentiles respectively.

NP-CI and BCa-CI were always fairly accurate with respect to the central percentiles, and in quasi-normal samples (e.g. S1) they had comparable performance to P-CI although producing slightly wider intervals and thus being less conservative (22). For n ≤ 120, negligible difference was found in the boundaries returned by NP-CI and BCa-CI respectively. Of course, in other models of skewness (e.g. logistic, log-logistic or Weibull), the location of data especially in small sized samples may produce some more pronounced deviations due to the different weight that tailing gains in the distribution shape. However, we are confident that the non-parametric method is actually suitable in the common practice since it is readily computable by means of electronic spreadsheets (as for instance the one provided in Supplementary Material).

In conclusion, CI should be included for percentiles whenever they are intended as estimates of a population parameter or a criterion to be applied outside the sample where they were estimated. To this end the theoretical distribution underlying the data structure should be always investigated in order to facilitate the choice of the more accurate and conservative parametric CI. In this regard, a statistical test able to quantify the degree of deviation together with a normality plot should be always used in order to assess the compliance of data with the available methods. In the case that parametric methods were not applicable or available, NP-CI and BCa-CI should be equally trusted for the central percentiles, whereas for the extreme percentiles the choice should be based on careful evaluation of the degree of skewness and the density of data in the tails of the distribution.

Supplementary material

Work Examples



Finding out percentiles of the non-central t distribution

The most challenging part of computing the P-CI is finding the value of the non-central t distribution. Actually, this is unavailable through Microsoft Excel (unless third-part add-ins are used) as well as some statistical packages most familiar in the biomedical field (e.g. Medcalc). Fortunately, the “Keisan” (Casio Computers Co., Tokyo, Japan) offers high-precision on-line calculation from 6 up to 50 significant figures (https://keisan.casio.com/exec/system/1180573219). In Keisan the computation allows only λ ≥ 0 whereas the P-CI can require a negative value. To this concern, recalling that λ is a factor shifting and skewing the central t distribution away from its symmetry about 0, it is easy to understand how non-central t percentiles can be simply flipped and sign-inverted to satisfy symmetry in the negative λ. Hence, placing a = α/2 and b = 1 - α/2 are the percentiles for λ, then for - λ they turn into a’ = - (1 - a) and b’ = - (1 - b) and thus a’ = - (1 - α/2) and b’ = - (α/2). For instance, given λ = 3 and ν = 20 it yields a = 1.027 and b = 5.663, then for λ = - 3 it turns into a’ = - 5.663 and b’ = - 1.027.

Practical issues in calculating NP-CI for extreme quantiles

Because Bin(n;q) is finite, calculation for extreme percentiles boundaries may lead to unreal results in small samples. For instance, with N = 20 the index of the lower boundary for the 2.5th percentile and the 95% coverage probability is actually - 1 if data are approximately normal. Thus, in order to form this CI in the appropriate manner, one may choose to set arbitrarily the lower boundary to the least observed value in the sample. Of course this way it is cut off part of the actual coverage probability that is bound to be systematically smaller than the theoretical value unless the sample size is adequately large (N > 300 by calculation). To this end there are other methods for computing NP-CI (although not via simple electronic spreadsheet) that overcome this limitation especially for small sized samples (22, 33, 35).


[1] Conflicts of interest None declared.



Armitage P, Berry G, Matthews JNS. Statistical methods in medical research. 4th ed. Malden, MA: Blackwell Science; 2001. p. 817.


Fraser CG, Kallner A, Kenny D, Petersen PH. Introduction: strategies to set global quality specifications in laboratory medicine. Scand J Clin Lab Invest. 1999;59:477–8. https://doi.org/10.1080/00365519950185184


Sciacovelli L, O’Kane M, Skaik YA, Caciagli P, Pellegrini C, Da Rin G, et al. Quality Indicators in Laboratory Medicine: from theory to practice. Preliminary data from the IFCC Working Group Project “Laboratory Errors and Patient Safety”. Clin Chem Lab Med. 2011;49:835–44. https://doi.org/10.1515/CCLM.2011.128


Ozarda Y. Reference intervals: current status, recent developments and future considerations. Biochem Med (Zagreb). 2016;26:5–16. https://doi.org/10.11613/BM.2016.001


Sandoval Y, Apple FS. The global need to define normality: the 99th percentile value of cardiac troponin. Clin Chem. 2014;60:455–62. https://doi.org/10.1373/clinchem.2013.211706


Neyman J. Outline of a theory of statistical estimation based on the classical theory of probability. Philos Trans R Soc Lond. 1937;236:333–80. https://doi.org/10.1098/rsta.1937.0005


Simundic AM. Confidence interval. Biochem Med (Zagreb). 2008;2:154–61. https://doi.org/10.11613/BM.2008.015


Lehmann EL, Casella G. Theory of point estimation. 2nd ed. New York: Springer; 1998. p. 589.


Solberg HE. RefVal: a program implementing the recommendations of the International Federation of Clinical Chemistry on the statistical treatment of reference values. Comput Methods Programs Biomed. 1995;48:247–56. https://doi.org/10.1016/0169-2607(95)01697-X


Solberg HE. The theory of reference values Part 5. Statistical treatment of collected reference values. Determination of reference limits. J Clin Chem Clin Biochem. 1983;21:749–60.


Linnet K. Nonparametric estimation of reference intervals by simple and bootstrap-based procedures. Clin Chem. 2000;46:867–9.


Solberg HE. The IFCC recommendation on estimation of reference intervals. The RefVal program. Clin Chem Lab Med. 2004;42:710–4. https://doi.org/10.1515/CCLM.2004.121


Horn PS, Pesce AJ. Reference intervals: a user’s guide. Washington, DC: American Association for Clinical Chemistry Press; 2005. p. 123.


Clinical and Laboratory Standards Institute (CLSI). Defining, establishing and verifying reference intervals in the clinical laboratory. Approved guideline-third edition. CLSI document EP28-A3c. Wayne, PA: CLSI, 2008.


Geffré A, Concordet D, Braun JP, Trumel C. Reference Value Advisor: a new freeware set of macroinstructions to calculate reference intervals with Microsoft Excel. Vet Clin Pathol. 2011;40:107–12. https://doi.org/10.1111/j.1939-165X.2011.00287.x


Coskun A, Ceyhan E, Inal TC, Serteser M, Unsal I. The comparison of parametric and nonparametric bootstrap methods for reference interval computation in small sample size groups. Accredit Qual Assur. 2013;18:51–60. https://doi.org/10.1007/s00769-012-0948-5


Theodorsson E. Resampling methods in Microsoft Excel® for estimating reference intervals. Biochem Med (Zagreb). 2015;25:311–9. https://doi.org/10.11613/BM.2015.031


Chakraborti S, Li J. Confidence interval estimation of a normal percentile. Am Stat. 2007;61:331–6. https://doi.org/10.1198/000313007X244457


Lawless JF. Statistical models and methods for lifetime data. 2nd ed. Hoboken, N.J.: Wiley-Interscience; 2003. p. 630.


Phan LD, McCool JI. Exact confidence intervals for Weibull parameters and percentiles. Proc Inst Mech Eng O. 2009;223:387–94. https://doi.org/10.1243/1748006XJRR247


Yoshitaka I, Noriyuki S, Hiroaki S. Bootstrap methods for confidence intervals of percentiles from dataset containing nondetected observations using lognormal distribution. J Chemometr. 2006;20:68–75. https://doi.org/10.1002/cem.987


Steinberg SM, Davis CE. Distribution-free confidence intervals for quantiles in small samples. Commun Stat Theory Methods. 1985;14:979–90. https://doi.org/10.1080/03610928508805144


Hahn GJ, Meeker WQ. Statistical intervals: a guide for practitioners and researchers. Haboken, NJ: John Wiley & Sons Inc; 1991. p. 392. https://doi.org/https://doi.org/10.1002/9780470316771


Campbell MJ, Gardner MJ. Calculating confidence intervals for some non-parametric analyses. Br Med J (Clin Res Ed). 1988;296:1454–6. https://doi.org/10.1136/bmj.296.6634.1454


Efron B, Tibshirani R. Bootstrap methods for standard errors, confidence intervals, and other measures of statistical accuracy. Stat Sci. 1986;1:54–75. https://doi.org/10.1214/ss/1177013815


Jones HL. Investigating the properties of a sample mean by employing random subsample means. J Am Stat Assoc. 1956;51:54–83. https://doi.org/10.1080/01621459.1956.10501311


Efron B. Bootstrap methods: another look at the Jackknife. Ann Stat. 1979;7:1–26. https://doi.org/10.1214/aos/1176344552


Efron B. Better bootstrap confidence intervals. J Am Stat Assoc. 1987;82:171–85. https://doi.org/10.1080/01621459.1987.10478410


Diciccio TJ, Romano JP. A review of bootstrap confidence intervals. J R Stat Soc Series B Stat Methodol. 1988;50:338–54.


Baser O, Crown WH, Pollicino C. Guidelines for selecting among different types of bootstraps. Curr Med Res Opin. 2006;22:799–808. https://doi.org/10.1185/030079906X100230


Henderson AR. The bootstrap: a technique for data-driven statistics. Using computer-intensive analyses to explore experimental data. Clin Chim Acta. 2005;359:1–26. https://doi.org/10.1016/j.cccn.2005.04.002


Haukoos JS, Lewis RJ. Advanced statistics: bootstrapping confidence intervals for statistics with “difficult” distributions. Acad Emerg Med. 2005;12:360–5. https://doi.org/10.1197/j.aem.2004.11.018


Zieliński R, Zieliński W. Best exact nonparametric confidence intervals for quantiles. Statistics. 2005;39:67–71. https://doi.org/10.1080/02331880412331329854


Kyselý J. Coverage probability of bootstrap confidence intervals in heavy-tailed frequency models, with application to precipitation data. Theor Appl Climatol. 2010;101:345–61. https://doi.org/10.1007/s00704-009-0190-1


Hayter AJ. Simultaneous confidence intervals for several quantiles of an unknown distribution. Am Stat. 2014;68:56–62. https://doi.org/10.1080/00031305.2013.869259