Abstract
We propose a sequential design method aiming at the estimation of an extreme quantile based on a sample of binary data corresponding to peaks over a given threshold. This study is motivated by an industrial challenge in material reliability and consists of estimating a failure quantile from trials whose outcomes are reduced to indicators of whether the specimen has failed at the tested stress levels. The proposed approach relies on a splitting strategy that decomposes the target extreme probability into a product of higher-order conditional probabilities, enabling a progressive exploration of the tail of the distribution through sampling under truncated laws. We consider GEV and Weibull models for the underlying distribution, and the sequential estimation of their parameters is carried out using an enhanced maximum likelihood procedure specifically adapted to binary data, addressing the substantial uncertainty inherent to such limited information.
Keywords: extreme quantile estimation, sequential design, binary information, splitting, extreme value theory
1. Introduction
Consider a non-negative random variable X with distribution function G. Let be n independent copies of The aim of this paper is to estimate , the -quantile of G when is much smaller than We therefore aim at the estimation of so-called extreme quantiles. This question has been handled by various authors, and these results are reviewed in Section 3. The approach that we develop is quite different since we do not assume that the ’s can be observed. For any threshold x, we define the following random variable:
which therefore has a Bernoulli distribution with parameter The threshold x is chosen by the experimenter, but the underlying value of X remains unobserved, leading to a substantial loss of information. Such settings are common in industrial statistics. For instance, when assessing the strength of a material or a component, one applies a load x and records only whether failure occurs.
In the following, we denote by R the resistance of this material and observe the corresponding indicator Y. Inference on G can be performed for large n, making use of many thresholds Unfortunately, such a procedure will not be of any help for extreme quantiles.
Indeed, this setting raises significant methodological challenges. First, the observation of binary responses only leads to a substantial loss of information compared to standard settings where the variable of interest is fully observed. Second, the estimation of extreme quantiles, associated with very small probability levels, further exacerbates this difficulty.
Existing approaches for quantile estimation from binary data, in particular those based on adaptations of the Robbins–Monro procedure, typically rely on a precise prior specification of the distribution of the latent variable, including in the vicinity of the target quantile. Such assumptions may be difficult to justify in practice, especially in industrial contexts where prior information is limited. Moreover, these methods are generally designed for moderate probability levels and do not naturally extend to the estimation of extreme quantiles under severe information constraints.
This motivates the development of alternative approaches that are able to explore the tail of the distribution in a progressive manner while reducing the reliance on strong prior assumptions.
To address these limitations, we will consider a design of experiments that progressively explores increasingly extreme regions of the distribution. More precisely, we assume that observations may be collected not only when R follows G, but also when R follows the conditional law of R given . Under such an assumption, it becomes possible to estimate even when , where n denotes the total number of trials.
In materials science, this amounts to considering trials based on artificially modified materials; when estimating extreme upper quantiles, this corresponds to strengthening the material. We would consider a family of increasing thresholds and for each of them realize trials, each block of i.i.d. realizations Y’s, therefore, corresponds to functions of unobserved R’s with distribution G conditioned upon ,
This setting departs significantly from classical approaches based on full data, and is particularly suited for industrial statistics and reliability studies in materials science. From a statistical viewpoint, the situation is tractable when G belongs to a parametric family for which the conditional law of R given preserves the functional form of G, up to changes in the parameters. In this case, sampling under conditional laws may be carried out adaptively through a sequential choice of the thresholds ’s. This leads to a recursive estimation procedure in which the parameters of G are updated iteratively, and the target quantile is obtained by combining quantiles of conditional distributions. This splitting-type approach guides the selection of the ’s so that can ultimately be determined from the final conditional distribution associated with threshold .
Such techniques are closely connected to safety analysis and pharmaceutical dose-finding, where interest naturally focuses on the behavior of the system under very small probability levels. Through a simple change of variable, these situations can be reframed in terms of upper-tail events. In particular, if , then for , the event is equivalent to . Accordingly, we make use of this duality throughout the paper, exploiting standard results for when necessary. In particular, if denotes the -quantile of R and the -quantile of , then . While this notation may appear slightly cumbersome, it is natural in industrial statistics.
This article is organized as follows, and its main contributions are introduced along the way. Section 2 formalizes the problem in the framework of an industrial application and highlights the specific challenges arising from binary observations and extreme quantile estimation. Section 3 reviews the existing literature on extreme quantile estimation and experimental design under binary data. This discussion emphasizes the limitations of current approaches in the regime of very small probabilities and motivates the need for alternative strategies. In Section 4, we introduce a novel splitting-based framework for extreme quantile estimation under partial observability. This constitutes a central contribution of the paper, as it enables the progressive exploration of the tail of the distribution through a sequential experimental design. The proposed procedure is elaborated for a Generalized Pareto model. Section 5 develops the associated estimation procedure. In particular, we propose a dual-criterion approach combining a likelihood-based component with a stability condition on sequentially estimated quantiles. Section 6 extends the methodology to an alternative parametric model, illustrating the flexibility of the proposed framework beyond the Generalized Pareto setting. Finally, Section 7 and Section 8 provide a brief discussion of model selection and behavior under misspecification, as well as hints about extensions of the models studied beforehand.
2. Problem Formulation in an Industrial Reliability Framework
This study focuses on the estimation of extreme failure quantiles, a critical issue in industrial risk assessment. Such quantiles play a major role in engineering applications, particularly in the aeronautics industry, where they intervene directly in decisions pertaining to engine component dimensioning and the management of fatigue-related risks. For a detailed presentation of the industrial context related to material fatigue that motivated this study, see Broniatowski and Miranda (2019) [1].
In this context, the usual estimation procedures rely on data obtained from experimental trial campaigns, in which specimens are subjected to various stress levels and tested until failure or until the end of the trial. Such experimental campaigns can be extremely costly in certain industrial contexts, such as aeronautics, which severely limit both the total number of trials and the diversity of experimental conditions that can be explored.
The aim of this study is to introduce a new experimental design methodology for the estimation of extreme failure quantiles at a fixed target lifetime, under a very low risk level of the order of , while relying on a minimal number of trials.
Let S denote the stress level applied (in MPa).
Throughout the paper, we will denote R, a positive r.v. modeling the resistance of the material at the target lifetime and homogeneous to the stress.
Define as the failure quantile of probability level , the level of stress that guarantees that the risk of failure before the fixed lifetime does not exceed . Thus, is the -quantile of the distribution of R:
| (1) |
However, R is not directly observed during the experiments. The relevant information available to characterize R is limited to indicators of whether or not the specimen tested has failed at s before the end of the trial. Therefore, the relevant observations corresponding to a campaign of n trials are formed by a sample of variables with for
where is the stress applied on specimen
Note that the number of observations is constrained by industrial and financial considerations. Thus, is way lower than , and we are considering a quantile lying outside the sample range.
Although this work is motivated by an industrial application in materials science, similar statistical settings arise in other fields, such as broader reliability analysis or dose-finding studies in medical trials, where the goal is to estimate a maximum tolerated dose corresponding to a very small failure probability.
3. Extreme Quantile Estimation—A Short Survey
As seen above, estimating extreme failure quantiles raises two main issues: on the one hand, the estimation of an extreme quantile, and on the other hand, the need to conduct inference based on exceedances under thresholds. We provide here a brief overview of these two areas, bearing in mind that the literature on extreme quantile estimation typically assumes complete data or, at best, right-censored observations.
3.1. Extreme Quantiles Estimation Methods
Extreme quantile estimation in the univariate setting is widely covered in the literature when the variable of interest X is either completely or partially observed. The usual framework is the study of the -quantile of a r.v X, with very small .
The most classical case corresponds to the setting where is drawn from an n sample of observations . A distinction is usually made between high quantiles that lie within the sample range (see Weissman 1978 [2] and Dekkers et al., 1989 [3]) and extreme quantiles that fall outside the range of the observations (e.g., De Haan and Rootzén 1993 [4]). It is assumed that X belongs to the domain of attraction of an extreme value distribution. The tail index of the latter is then estimated through maximum likelihood (Weissman 1978 [2]) or through an extension of Hill’s estimator, such as the moment estimator by Dekkers et al. (1989) [3]. The estimator of the quantile is then deduced from the inverse function of the distribution of the k largest observations. Note that all the above references assume that the distribution has a Pareto tail.
Alternative modeling strategies have been proposed by De Valk (2016) [5] and De Valk and Cai (2018) [6], who assume a Weibull-type tail. This relaxes certain second-order hypotheses on the tail and allows one to target quantiles far outside the sample range. We will use these methods as benchmarks in our empirical comparisons.
Recent studies have also tackled the issue of censoring. For instance, Beirlant et al. (2007) [7] and Einmahl et al. (2008) [8] proposed a generalization of the peak-over-threshold method when the data are subjected to random right censoring and an estimator for extreme quantiles. The idea is to consider a consistent estimator of the tail index on the censored data and divide it by the proportion of censored observations in the tail. Worms and Worms (2014) [9] studied estimators of the extremal index based on Kaplan–Meier integration and censored regression.
By contrast, the literature is much sparser in the case of complete truncation, i.e., when only exceedances over predetermined thresholds are observed. All of the methods mentioned above rely on the higher-order statistics of the original sample, which are not available in the present setting. Consequently, classical extreme quantile estimators are not directly applicable to our problem.
In the next subsection, we therefore turn to sequential experimental designs used in industrial and biostatistical contexts, and assess their relevance for estimating extreme quantiles from dichotomous data.
3.2. Sequential Design Based on Dichotomous Data
A few studies have addressed the estimation of quantiles from binary data. Wu (1985) [10] and Joseph (2004) [11] propose adaptations of the Robbins and Monro procedure for binary data. Yet both focus on quantiles of moderate order (typically between 0.1 and 0.9) and require a fairly accurate prior knowledge of the latent distribution of the variable of interest around the targeted quantile. Such an assumption may be unrealistic in many applications, especially when targeting the tail of the distribution.
Wu and Tian (2014) [12] introduced what appears to be the most advanced procedure for quantile estimation based on binary data. Their three-step sequential approach (search, estimate, approximate) shows promising performance even for extreme quantiles, but it relies heavily on the specifics of the application context and must be tailored on a case-by-case basis.
In contrast, the approach we propose in this study aims to provide a generic framework specifically designed for the estimation of extreme failure quantiles. Our methodology is grounded in parametric assumptions justified by extreme value theory results, thus reducing the need for strong prior knowledge on the latent distribution and reducing sensitivity to model specification.
The remainder of this section reviews two classical designs frequently used in industry and biostatistics and conceptually closest to our objective: the Staircase and the Continual Reassessment Method (CRM). In what follows, we consider the estimation of a small quantile for events of the form with small. Both procedures rely on a parametric model for the strength variable R. We retain two specifications to facilitate performance comparisons, rather than to provide fully realistic safety assessments.
3.2.1. The Staircase Method
Denote . Invented by Dixon and Mood (1948) [13] and refined in Dixon (1965) [14], this technique aims at the estimation of the parameter through sequential search based on data of exceedances under thresholds. Starting from an initial stress level , each item is tested and the next test level is adjusted by a fixed increment : it is increased after a survival and decreased after a failure. This process is repeated for the K specimens. After the K trials, the parameter is estimated through maximization of the likelihood.
The proper conduct of the Staircase method relies on strong assumptions about the choice of the design parameters. Firstly, has to be sufficiently close to the expectation of R and secondly, has to lie between and , where designates the standard deviation of the distribution of R.
Numerical Results
The accuracy of the procedure has been evaluated on the two models presented below on a batch of 1000 replications, each with
Exponential case
Let with . The input parameters are and .
As shown in Table 1, the relative error pertaining to the parameter is roughly , although the input parameters are reasonably well chosen for the method. The resulting relative error on the quantile is Indeed, the parameter is underestimated, which results in an overestimation of the variance , which induces an overestimation of the quantile.
Table 1.
Results obtained using the Staircase method through simulations under the exponential model.
| Relative Error | |||
|---|---|---|---|
| On the Parameter | On | ||
| Mean | Std | Mean | Std |
| −0.252 | 0.178 | 0.4064874 | 0.304 |
Gaussian case
We now choose with and . The value of is set to the expectation and belongs to the interval The same procedure as above is performed and yields the results in Table 2.
Table 2.
Results obtained using the Staircase method through simulations under the Gaussian model.
| Relative Error | |||||
|---|---|---|---|---|---|
| On | On | On | |||
| Mean | Std | Mean | Std | Mean | Std |
| −0.059 | 0.034 | 1.544 | 0.903 | −1.753 | 0.983 |
The expectation of R is recovered rather accurately, whereas the estimation of the standard deviation suffers a loss in accuracy, which, in turn, yields a relative error of 180 % on the quantile.
Limitations of the Staircase Method
While the Staircase method can recover the central tendency with a limited number of trials, it is not suitable for extreme quantile estimation. The latter relies on extrapolation from potentially biased parameter estimates, and simple reparametrizations (e.g., in terms of the extreme quantile) do not remedy the inherent loss of accuracy.
3.2.2. The Continuous Reassessment Method (CRM)
General Principle
The CRM (O’Quigley, Pepe and Fisher, 1990 [15]) has been designed for clinical trials to estimate among J stress levels , when is of order .
Denote . The estimator of is
This optimization is performed iteratively with K trials per iteration. Starting with an initial estimator of , for example, through a Bayesian choice as proposed in [15], define
Each iteration follows a two-step procedure:
Update using all past data (maximum likelihood or Bayesian posterior under ).
- Set
and perform the next K Bernoulli trials at level .
The stopping rule depends on the context (maximum number of trials or stabilization of the results).
Note that the Bayesian inference is useful in cases where there is no diversity in the observations at some iterations of the procedure, i.e., when, at a given level of test , only failures or survivals are observed.
Application to Failure Quantiles
Denote by the prior indexed by the stress level s. models the failure probability at level s and has a Beta distribution given by
| (2) |
It amounts to stating that at a given stress level s, we expect k failures out of n trials.
Let R follow an exponential distribution: .
It follows .
Define the random variable which, by definition of , is distributed as a k-order statistic of a uniform distribution .
The estimation procedure of the CRM is obtained as follows:
Step 1. Compute an initial estimator of the parameter
with . Define
and perform J trials at level . Denote the observations
Step i. At iteration i, compute the posterior distribution of the parameter:
| (3) |
The above distribution also corresponds to an order statistic of the uniform distribution . We then obtain an estimate .
The next stress level to be tested in the procedure is then given by
Numerical Simulation for the CRM
Under the exponential model with parameter , using iterations, equally spaced thresholds , and trials per iteration, we obtain the results reported in Table 3.
Table 3.
Results obtained through CRM on simulations for the exponential model.
| Relative Error | |||
|---|---|---|---|
| On the -Quantile | On the -Quantile | ||
| Mean | Std | Mean | Std |
| 0.129 | 0.48 | −0.799 | 0.606 |
The -quantile is poorly estimated even in this simple setting. Near the target threshold, almost no failures are observed; for acceptable values of K, the method becomes ineffective. Figure 1 illustrates the improvement in accuracy as K increases.
Figure 1.
Relative error on the -quantile with respect to the number of trials for each stress level.
In summary, both the Staircase and the CRM face the same limitation for extreme quantile estimation. The staircase targets the central tendency, whereas the CRM is calibrated for moderate quantile levels (around ), far from the target . This motivates the original procedure proposed in the next sections, specifically designed for extreme quantiles under binary information.
4. A New Design for the Estimation of Extreme Quantiles
4.1. Splitting
The design we propose is directly inspired by the general principle of splitting methods used in the domain of rare events simulation and introduced by Kahn and Harris (1951) [16].
The central idea is to overcome the difficulty of estimating an extremely small probability by decomposing the target event into a sequence of events of higher probability. Splitting achieves this by expressing a rare-event probability as a product of conditional probabilities that are easier to estimate.
Let denote the distribution of the random variable R. The event can be expressed as the intersection of inclusive events for it holds:
It follows that
| (4) |
The thresholds should be chosen so that each conditional probability is of moderate level (typically or ). This ensures that the event is observable under the conditional distribution of , while the product of these probabilities still reconstructs the target rare-event probability via (4), with a small number of stages m.
From the formal decomposition in (4), a practical experimental scheme can be deduced. Its structure is given in Procedure 1.
Remark 1.
Our approach can be seen as an alternative to stochastic approximation methods for quantile estimation under binary responses. While these methods rely on recursive updates driven by observed responses, they typically require either moderate quantile levels or strong prior assumptions on the underlying distribution.
In contrast, the procedure proposed here combines a sequential design with structural modeling inspired by extreme value theory, allowing the exploration of much more extreme regions of the distribution under weaker assumptions.
Remark 2.
Practical feasibility: The above procedure relies on the assumption that sampling can be performed from the conditional distribution of the resistance at each step. In practice, this amounts to conducting experiments on specimens with progressively reduced resistance levels. Although this assumption may appear strong, it is, in fact, realistic in some experimental settings, as specimens with controlled weakened resistance can be manufactured through appropriate machining processes. An illustrated example of such an approach in the context of material fatigue is provided in Broniatowski and Miranda (2019) [1].
| Procedure 1 Splitting procedure |
|
4.2. Choice of Sequential Design Parameters
The performance of the proposed sequential procedure depends on several parameters, namely the conditional probability level p, the number of stages m, and the number of trials per stage K. Their selection involves a trade-off between statistical accuracy and experimental cost.
A standard choice for the conditional probability level p is to fix it around . This value ensures that the conditional failure probability remains sufficiently large to be reliably estimated from a limited number of binary observations, while still enabling a progressive exploration of the tail of the distribution through the splitting mechanism.
The first threshold is selected so as to approximate the target conditional probability level p. In practice, however, the resulting probability may differ from p. In such cases, the model parameters are updated to ensure consistency with the target rare event probability , typically through the relation
The selection of may also rely on domain-specific expertise, depending on the application context. In principle, the first iteration could be refined by incorporating additional prior information, for instance, regarding the central tendency of the distribution. However, such extensions are not considered in the present work in order to preserve the generality of the proposed methodology.
More generally, the parameters p, m, and K must be chosen jointly. Increasing the number of stages m allows for a more gradual progression toward extreme regions, which may improve robustness and stability, but also increases the overall complexity of the procedure. Increasing the number of trials K at each stage improves the accuracy of the parameter estimates and of the conditional quantiles, at the cost of a higher experimental burden. Finally, the choice of p controls the balance between these two effects: smaller values of p lead to a faster progression toward the tail but reduce the amount of information available at each stage, whereas larger values improve estimation reliability but require more stages to reach the target rare event probability.
In practice, this trade-off is largely driven by application-specific constraints. In industrial settings, for instance, the cost of performing experiments may increase significantly as testing conditions become more severe (corresponding to deeper levels in the sequential procedure), while the availability of resources (time, equipment, personnel) may limit the number of feasible trials. As a result, the parameters p, m, and K are often not freely chosen by the statistician, but rather dictated by these operational constraints. The guidelines provided here should therefore be understood as indicative rather than prescriptive. A detailed illustration of these trade-offs in an industrial context is provided in Broniatowski and Miranda (2019) [1].
4.3. Modeling the Distribution of the Strength, Pareto Model
The events under consideration have a small probability under By (4), we are led to consider the limit behavior of conditional distributions under smaller and smaller thresholds. To this end, we rely on the classical approximations due to Balkema and de Haan (1974) [17] and Pickands (1975) [18] stated below in the familiar setting of exceedances above high thresholds. Let .
Theorem 1.
For of distribution F belonging to the maximum domain of attraction of an extreme value distribution with tail index c, i.e., , it holds that: there exists , such that:
where is defined through
where and .
The distribution G is the Generalized Pareto distribution given explicitly by
where for and if
A key feature of GPDs is their invariance under threshold conditioning. Indeed it holds, for and ,
| (5) |
We therefore state:
Proposition 1.
When then, given , the random variable follows a .
Thus, GPDs are both stable under thresholding and arise as limiting models for threshold exceedances. This mirrors the classical rationale that motivates the use of normal or stable laws in additive models. These properties make GPDs particularly suitable to model for excess-probability inference. Due to the lack of memory property, the exponential distribution, which appears as a possible limit distribution for excess probabilities when in Theorem 1, does not qualify for modeling. Moreover, since R may take arbitrarily small values (i.e., is unbounded), we restrict attention to .
Turning to the context of extreme failure quantiles, we make use of the random variable and proceed to the corresponding change of variable.
When , the distribution function of the r.v. R writes for nonnegative x:
| (6) |
For , the conditional distribution of R given is
showing that the distribution of R is stable under threshold conditioning with parameter , where
| (7) |
In practice, at step j of the procedure, the stress level is the corresponding threshold , where is a right-quantile of the conditional law of given . Therefore, the observations take the form .
A convenient feature of model (6) lies in the fact that the conditional distributions are completely determined by the initial distribution of R, therefore by and The parameters of the conditional distributions are determined from these initial parameters and by the corresponding stress level see (7).
4.4. Notation and Framework for Sequential GPD Modeling and Parameterization
We assume that the distribution of the random variable follows a Generalized Pareto Distribution with distribution function , and denote .
The proposed procedure relies on a sequence of increasing thresholds and exploits the stability property of the GPD under threshold exceedances. As stated in (5), for each , the conditional distribution of given remains a GPD and can be written as
with and .
At iteration j, we denote by the estimators of , so that provides an estimate of . The parameters of the initial distribution can then be recovered from through
4.5. Sequential Design for the Extreme Quantile Estimation
Fix m and p, where m denotes the number of stress levels under which the trials will be performed, and p is such that
Select an initial stress level that is sufficiently large (i.e., sufficiently small) so that is large enough and perform trials at this level. Ideally, one would choose such that , which cannot be secured; in practice, this choice relies on expert judgment.
Switch to the transformed variable . Estimate the GPD parameters of based on the observations above , producing the initial estimates . As stated in Section 4.2, is chosen so that it corresponds to medium stress conditions, and failures are therefore easy to observe experimentally.
Define
the -quantile of is the level of stress to be tested at the following iteration.
Iterating from step to , perform K trials under say and consider the observable variables . Therefore, the K iid replications follow a Bernoulli , where has been determined at the previous step of the procedure. Estimate in the resulting Bernoulli scheme, say . Then define
which is the -quantile of the estimated conditional distribution of given , i.e., , and the next level to be tested.
In practice a conservative choice for m is , where denotes the ceiling function. The attained probability is a proxy of (see Section 5).
The m stress levels satisfy
Finally, by construction, is a proxy of .
Although conceptually simple, the method raises several challenges, chiefly regarding the estimation of at each stage. The next section addresses this issue.
5. Enhanced Sequential Design in the Pareto Model
In this section, we focus on the estimation of the parameters in the distribution of One of the main difficulties lies in the fact that the available information does not consist of replications of the random variable under the current conditional distribution of given but merely on the very downgraded functions of those.
At step j, we are given and derive as its -quantile. Simulating K random variable with distribution , the only observable quantities are the Bernoulli(p) variables
which represent a substantial loss of information compared with the underlying . Estimating from these ’s is therefore intrinsically challenging.
5.1. Limitations of Likelihood-Based Estimation
A natural approach is to estimate the parameters at each iteration using maximum likelihood. However, this leads to a rapid deterioration of the estimation of the extreme quantile for small . In simulation experiments, the large standard deviation of is directly linked to the poor precision of the iterative estimators . To illustrate this, we generated independent Bernoulli variables with parameter . Figure 2 displays the corresponding log-likelihood profile as the Bernoulli parameter varies with . As expected, the log-likelihood surface is extremely flat across a wide region of the parameter space.
Figure 2.
Log-likelihood of the Pareto model with binary data.
This explains the poor results in Table 4 obtained through the Splitting procedure when the parameters at each step are estimated by maximum likelihood, particularly the high variability in the estimated quantiles. Moreover, the accuracy of the estimator of quickly decreases with the number K of replications , , as illustrated by results in Table 5.
Table 4.
Estimation of the -quantile, , through procedure Section 4.5 with .
| Minimum | Q25 | Q50 | Mean | Q75 | Maximum |
|---|---|---|---|---|---|
| 67.07 | 226.50 | 327.40 | 441.60 | 498.90 | 10,320.00 |
Table 5.
Estimation of the -quantile, , through procedure Section 4.5 for different values of K.
| for | for | |||
|---|---|---|---|---|
| Mean | Std | Mean | Std | |
| 469.103 | 1276.00 | 12,576.98 | 441.643 | 562.757 |
Changing the estimation criterion by some alternative method does not improve significantly; Figure 3 shows the distribution of the resulting estimators of for various estimation methods (minimum Kullback–Leibler, minimum Hellinger, and minimum L1 distances) of
Figure 3.
Estimations of the -quantile based on the Kullback–Leibler, L1 distance, and Hellinger distance criterion.
These observations motivate the development of an enhanced estimation strategy.
5.2. An Enhanced Sequential Criterion for Estimation
To overcome the limitations of standard estimation methods, we introduce an additional estimation criterion that exploits the iterative structure of the procedure. The main idea is to enforce stability by ensuring coherence between successive conditional quantile estimates through a backward validation mechanism.
Due to its backwards nature, this estimation procedure stands for iterations .
At iteration , the sample , is collected in the form of binary responses
where are drawn from . The empirical failure probability given by
| (8) |
provides an estimate of and a proxy of p. The conditional probability can be written as a function of the estimated parameters obtained at iteration j, namely using (7). This relies on the stability of the GPD under threshold exceedances, which ensures that parameters estimated at level j remain consistent with the distribution at previous levels.
At step estimate making use of This backward estimator writes
The distance
| (9) |
should be small, since both and should approximate
Consider the distance between quantiles
| (10) |
An estimate can be proposed as the minimizer of the above expression for for all j. This backward estimation provides coherence with respect to the unknown initial distribution through a retroactive validation of the parameters. If we had started with a good guess , then the successive , etc., would make (10) small, since (resp. ) would estimate the -conditional quantile of (resp. ).
It remains to restrict the minimization of (10) to a set of plausible parameter values. We suggest constructing a confidence region for the parameter With defined in (8) and , define the -confidence region for p by
| (11) |
where is the -quantile of the standard normal distribution. Define
| (12) |
Therefore, is a plausible set for
We summarize this discussion:
At iteration the estimator of is a solution of the minimization problem
| (13) |
where optimization is performed using standard numerical methods.
This backward estimation procedure enforces coherence with respect to the unknown initial distribution by ensuring consistency between successive conditional quantile estimates. If the initial parameters were equal to the true values , then the successive thresholds would accurately track the corresponding conditional quantiles, making the discrepancy in (10) negligible.
This criterion can be interpreted as a backward validation mechanism, in the spirit of cross-validation or jackknife-type approaches, where estimates at a given stage are assessed through their ability to reproduce quantities observed at previous stages. From a statistical perspective, it reflects the sequential nature of the data acquisition process by incorporating past information into the current estimation step, thereby ensuring coherence between successive conditional quantile estimates. It should therefore not be viewed as a likelihood-based or regularization approach, but rather as a consistency-based validation criterion embedded within the sequential design.
5.3. Theoretical Insight into the Estimator
Although a full theoretical analysis of the proposed estimator is beyond the scope of this paper, some insight into its behavior can be provided. The method relies on a splitting strategy that decomposes a rare event into a sequence of conditional events, thereby progressively exploring the tail of the distribution. Each step involves the estimation of an intermediate conditional quantile, and the final extreme quantile is obtained by the composition of these levels.
From an extreme value theory perspective, this construction can be interpreted as iteratively approximating increasingly extreme quantiles of the original distribution. As the number of iterations increases, the procedure effectively focuses on the tail region, where classical asymptotic results for conditional excess distributions apply.
Moreover, the recursive structure of the algorithm contributes to its stability: provided that the parameter estimates at a given step are sufficiently accurate, the subsequent sampling is concentrated in a relevant region of the distribution. The enhanced estimation criterion introduced in Section 5 further reinforces this stability by promoting consistency between successive conditional quantile estimates. From a heuristic perspective, the sequential updating mechanism may be interpreted as a fixed-point-type iteration, where parameter estimates are adjusted to ensure consistency with previously observed conditional quantiles. This suggests that convergence may be expected under suitable conditions, although no formal result is established in the present work. This reasoning relies on asymptotic arguments from extreme value theory, in particular on the applicability of the De Haan theorem, which ensures that the tail behavior can be approximated by a generalized Pareto distribution. Overall, these elements provide intuition on the expected convergence behavior of the method, even though a formal proof of consistency remains an open question.
5.4. Simulation-Based Numerical Results
This procedure has been applied in three cases. The reference case is . The second case, , corresponds to a lighter tail than the reference. The third case, , corresponds to a distribution with the same tail index as the reference but a larger dispersion index.
The overall procedure can be summarized as follows in Procedure 2:
Performance is primarily assessed using relative error, which provides a scale-invariant measure particularly suited to extreme quantile estimation, where the quantities of interest may span several orders of magnitude. From a practical standpoint, this metric also reflects the ability of the method to recover the correct order of magnitude, which is often the main objective in industrial applications.
Table 6 highlights important features of the proposed estimator. First, the estimation of deteriorates as the tail of the distribution becomes heavier, reflecting the intrinsic difficulty of extrapolating further into the tail. In addition, the estimator tends to underestimate , implying a conservative nature of the sequential exploration of extreme regions.
Table 6.
Mean and std of relative errors on the -quantile of GPD calculated through 400 replicas of Procedure 2.
| Parameters | Relative Error on | |
|---|---|---|
| Mean | Std | |
| , and | −0.222 | 0.554 |
| , and | −0.504 | 0.720 |
| , and | 0.310 | 0.590 |
Despite these limitations, the proposed method shows a clear improvement over the simple Maximum Likelihood estimation. This gain is particularly pronounced in the case of heavy-tailed distributions, where classical approaches struggle to provide reliable estimates Figure 4.
| Procedure 2 Simulation and dual-criterion estimation procedure | ||||
|
Figure 4.
Estimations of the -quantile of two GPD obtained by Maximum Likelihood and by the improved Maximum Likelihood method. The red line stands for the real value of .
Moreover, the influence of the number K of replications at each step reveals an important advantage of the method. While reducing K naturally increases variability, it also amplifies the relative gain over Maximum Likelihood estimation. This behavior, illustrated in Figure 5, suggests a certain robustness of the proposed approach in low-information settings, where only a limited number of trials can be performed at each stage.
Figure 5.
Estimations of the -quantile of a obtained by Maximum Likelihood and by the improved Maximum Likelihood method for different values of K. The red line stands for the real value of .
5.5. Performance of the Sequential Estimation
As stated in Section 5, a substantial amount of information is lost under complete truncation and binary sampling.
As stated in Section 3, most available approaches either assume fully observed data or rely on strong prior knowledge of the underlying distribution, particularly in the tail region. These assumptions are not compatible with the binary observation framework considered here. Thus, direct comparisons with existing methods for extreme quantile estimation are not straightforward in our setting.
For this reason, we instead compare our procedure to a benchmark based on full data observations, which provides an upper bound on the achievable performance under ideal information conditions.
5.5.1. Estimation of an Extreme Quantile Based on Complete Data, De Valk’s Estimator
In order to provide an upper bound for performance, we use the estimator proposed by De Valk and Cai (2018) [6]. Their framework targets quantiles of order with , where n is the sample size, which aligns with the industrial context that motivated this work. De Valk’s proposal is a modified Hill estimator adapted to log-Weibull-tailed models. De Valk’s estimator is consistent, asymptotically normally distributed, though it may exhibit finite-sample bias.
We briefly recall the assumptions underpinning De Valk’s approach. Let be n iid r.v’s with distribution F, and denote the order statistics. A tail regularity assumption is needed in order to estimate a quantile with order greater than
Denote , and let the function q be defined by
for .
Assume that
| (14) |
where g is a regularly varying function and
de Valk writes condition (14) as .
Despite its naming of log-Generalized tails, this condition also holds for Pareto-tailed distributions, as can be checked, providing
We now introduce de Valk’s extreme quantile estimator.
Let
Let be the quantile of order of the distribution F. The estimator makes use of , an intermediate order statistic of , where tends to infinity as and
de Valk’s estimator writes
| (15) |
When the support of F overlaps , then the sample size n should be large; see de Valk [6] for details.
Note that, in the case of a , parameter is known and equal to 1, and the normalizing function g is defined by for .
5.5.2. Loss in Accuracy Due to Binary Sampling
Table 7 compares the performance of De Valk’s method (using complete data) with that of our sequential procedure (using only dichotomous outcomes). Unsurprisingly, de Valk’s estimator outperforms ours. This gap reflects the loss of information induced by thresholding and dichotomization. Nevertheless, the comparison remains informative: even though our estimator typically exhibits a larger bias, its dispersion is of the same order of magnitude when handling heavy-tailed GPD models. Given the binary nature of the data, the average relative error is quite honorable. Overall, we can assess that a large part of the volatility of the estimator produced by our sequential methodology is due to the nature of the GPD model, as well as to the sample size.
Table 7.
Mean and std of the relative errors on the -quantile of GPD on complete and binary data for samples of size , computed through 400 replicas of both estimation procedures. Estimates on complete data are obtained with de Valk’s method; estimates on binary data are provided by the sequential design.
| Relative Error on the -Quantile | ||||
|---|---|---|---|---|
| Parameters | on Complete Data | on Binary Data | ||
| Mean | Std | Mean | Std | |
| , and | 0.052 | 0.257 | −0.222 | 0.554 |
| , and | 0.086 | 0.530 | −0.504 | 0.720 |
| , and | 0.116 | 0.625 | 0.310 | 0.590 |
6. Sequential Design for the Weibull Model
The main property that led to the GPD model is its relevance for tail modeling and its stability through threshold conditioning. In this section, we also investigate a parameterization based on the Weibull distribution as one of the most classical and widely used models in reliability engineering. Under this assumption, the conditional distribution of given takes a rather simple form, which allows for some variation of the sequential design method.
6.1. The Weibull Model
Let with scale and shape , and denote by G its c.d.f., g its density, and the quantile function. For ,
The conditional distribution of given truncation above a threshold is a truncated Weibull.
Let denote the distribution function of given .
A useful identity follows. For ,
| (16) |
Assuming and given , we may find the conditional quantile of order of the distribution of given . This solves the first step of the sequential estimation procedure through
where the parameter has to be estimated on the first run of trials.
The same logic extends iteratively. For
| (17) |
At iteration j, the thresholds and are known; the threshold is the -quantile of the conditional distribution, , hence solving
where the estimate of is updated from the data collected at iteration
6.2. Numerical Results
In line with Section 5.4 and Section 5.5, we assess the performance of the sequential design under a Weibull model. We estimate the -quantile in three scenarios. In the first case, the scale parameter a and the shape parameter b satisfy . This corresponds to a strictly decreasing density function with a heavy tail. In the second case, , the distribution is more skewed; in the third, , the distribution is less dispersed with a lighter tail.
Table 8 shows that the performance of our procedure is, as expected, sensitive to the shape of the distribution. The estimators are less accurate in case 1, corresponding to a heavier tail. We compare these errors to those obtained with complete data using de Valk’s methodology. The loss of accuracy due to data deterioration (binary sampling and truncation) is similar to that observed in the Pareto case, though slightly more pronounced. This is consistent with the fact that the Weibull family is less naturally aligned with the splitting structure than the GPD, which enjoys exact stability under threshold conditioning.
Table 8.
Mean and std of relative errors on the -quantile of Weibull distributions on complete and binary data for samples of size computed through 400 replicas. Estimates on complete data are obtained with de Valk’s method; estimates on binary data are provided by the sequential design.
| Relative Error on the -Quantile | ||||
|---|---|---|---|---|
| Parameters | on Binary Data | on Complete Data | ||
| Mean | Std | Mean | Std | |
| , et | 0.282 | 0.520 | 0.127 | 0.197 |
| , et | −0.260 | 0.490 | 0.084 | 0.122 |
| , et | −0.241 | 0.450 | 0.088 | 0.140 |
7. Model Selection and Misspecification
In the above sections, we considered two models primarily motivated by their theoretical properties. As stated in Section 4.3, modeling by a with is justified when the support of the original variable R may be bounded by 0. Note, however, that the GPD model naturally extends to , which corresponds to the exponential distribution and represents a trivial limiting case for excess probabilities.
Although we excluded the exponential case when modeling the excess probabilities of with a GPD, we nevertheless considered the Weibull model in Section 6, which lies in the maximum domain of attraction associated with . Beyond its compatibility with the splitting structure, the Weibull law is a classical and widely used model in reliability, which makes it a natural candidate for an adapted version of our sequential method.
In this section, we discuss the modeling decisions and give some hints on how to deal with misspecification.
7.1. Model Selection
The decision between a Pareto-type model with a strictly positive tail index and a Weibull-type model has been addressed in the literature; a variety of tests exist for assessing membership to a given maximum domain of attraction.
Dietrich et al. (2002) [19] Drees et al. (2006) [20] both propose a test for extreme value conditions related to Cramer-von Mises tests. Let X of distribution function G. The null hypothesis is
In our case, the theoretical value for the tail index is . The former test builds on the tail empirical quantile function, while the latter relies on a weighted approximation of the tail empirical distribution.
Choulakian and Stephens (2001) [21] propose a goodness-of-fit test in the spirit of Cramér–von Mises statistics, where unknown parameters are replaced by maximum likelihood estimates. Given a sample from G, the hypothesis to be tested is : the sample stems from a . The associated test statistics are given by the following equation:
where denotes the -th order statistic of the sample. The authors provide the corresponding tables of critical points.
Jurečková and Picek (2001) [22] designed a non-parametric test for determining whether a distribution G is light or heavy-tailed. The null hypothesis is defined by the following:
with fixed hypothetical . The test procedure consists of splitting the data set into N samples and computing the empirical distribution of the extrema of each sample.
The practical assessment of model suitability in our industrial context is delicate. Two main obstacles arise: (i) data are collected sequentially during the procedure, with no upfront sample of available observations; (ii) the variable of interest R is not observed directly and only dichotomized exceedances over chosen thresholds are available. Most existing procedures assume fully observed data and are semi- or nonparametric tests built from order statistics; their performance also relies on large samples, which are unrealistic in our setting due to the cost and duration of experimental trials.
As an alternative, one may resort to a posteriori validation, once the procedure is completed, combining the statistical output with expert judgment to confirm (or reject) the chosen model.
7.2. Misspecification
In Section 4.3, we assumed that follows a GPD at the outset. In practice, the tail of the true distribution may converge toward a GPD as thresholds increase, yet differ from it at finite levels.
7.2.1. Robustness to Model Misspecification
The proposed approach relies on parametric assumptions, in particular the generalized Pareto distribution, which is motivated by its stability under threshold exceedances in extreme value theory. This choice is especially relevant in the later stages of the splitting procedure, where the algorithm focuses on increasingly extreme regions of the distribution.
It should be noted, however, that the parametric assumption may be less accurate during the early iterations, when the data are not yet in the tail regime. Nevertheless, the sequential nature of the method progressively reduces the influence of these initial stages, as the final estimate is primarily driven by the deepest levels of the splitting procedure.
This behavior is consistent with standard extreme value estimation techniques, which rely primarily on the largest observations and are therefore relatively insensitive to the distribution in central regions. As a result, the method is expected to exhibit some degree of robustness to model misspecification, provided that the tail behavior is adequately captured.
A more systematic assessment of this robustness, for instance, through simulation studies under alternative distributions, would be a valuable direction for future work. However, such an analysis is not straightforward in the present sequential setting, where data are generated adaptively under conditional distributions. In particular, standard tools such as influence functions are difficult to implement, and a comprehensive empirical validation would go beyond the scope of the present work.
7.2.2. Insights on Control of Model Deviation
While the previous subsection discusses the robustness of the proposed approach from an asymptotic perspective, it is also important to quantify the effect of model misspecification in finite-sample settings, especially in the early iterations of the procedure, where the extreme regime has not yet been reached.
In this subsection, we address this issue by introducing a neighborhood-based framework that allows us to control deviations from the generalized Pareto model and to guide the selection of appropriate threshold levels within the sequential design.
Let us assume that does not follow a GPD with distribution function F, but rather a distribution G whose tail becomes increasingly close to that of a GPD.
In this case, the issue is to control the distance between G and the theoretical GPD and to determine a threshold level beyond which this discrepancy becomes negligible. One way to proceed is to restrict attention to a neighborhood of F:
| (18) |
where and w is an increasing weight function such that . defines a neighborhood, which does not tolerate large departures from F in the right tail of the distribution.
Let . It follows from (18) that a bound for the conditional distribution of x given is as follows:
| (19) |
When , the bounds of (19) match the conditional probabilities of the theoretical Pareto distribution.
In order to control the distance between F and G, this bound may be rewritten in terms of relative error with respect to the Pareto distribution. Using a first-order Taylor expansion for small , we obtain
| (20) |
where
For a given close to 0, the relative error on the conditional probabilities can be controlled through the choice of s. In particular, the relative error is bounded by a prescribed level whenever
This provides a practical guideline for selecting thresholds that mitigate the impact of misspecification within the sequential design.
8. Perspectives, Generalization of the Two Models
In this work, we have considered two models for that exploit the thresholding operations used in the splitting method. This choice is, however, restrictive: the limited information obtained from the trials prevents the use of highly flexible models for the distribution of the resistance. In the following, we outline several possible extensions and generalizations of those models based on common properties of the GPD and Weibull models.
8.1. Variations Around Mixture Forms
When the tail index is positive, the GPD is completely monotone, and thus can be written as the Laplace transform of a probability distribution. Thyrion (1964) [23] and Thorin (1977) [24] established that a , with , can be written as the Laplace transform of a Gamma random variable V whose parameters are functions of and : . Let v denote the density of V,
| (21) |
It follows that the conditional survival function of , , is given by the following:
with and .
Expression (21) gives room to an extension of the Pareto model. Indeed, we could consider distributions of that share the same mixture form with a mixing variable W that possesses some common characteristics with the Gamma-distributed random variable
Similarly, the Weibull distribution can also be written as the Laplace transform of a stable law of density g whenever . Indeed, it holds from Feller 1971 [25] (p. 450, Theorem 1) that:
| (22) |
where g is the density of an infinitely divisible probability distribution.
It follows that, for
| (23) |
Thus, an alternative modeling of could consist of any distribution that can be written as a Laplace transform of a stable law of density defined on and parametrized by , that satisfies the following property. For any , the distribution function of the conditional distribution of given can be expressed as the Laplace transform of where
where is defined in (23).
8.2. Variation Around the GPD
Another approach, inspired by Naveau et al. (2016) [26], is to modify the model so that the distribution of converges to a GPD in the upper tail while taking a more flexible form near 0.
is generated through with . Let us consider now a deformation of the uniform variable defined on , and the transform W of the GPD: .
Since the survival function of the GPD is completely monotone, we can choose W so that the distribution of retains this property.
Proposition 2.
If is completely monotone and ψ is a positive function whose derivative is completely monotone, then is completely monotone.
The transformation of the GPD has a cumulative distribution function and survival function . is a Bernstein function; thus, is completely monotone if is as well.
Examples of Admissible Functions
(1) Exponential form:
The obtained transformation is: ,
with completely monotone.
(2) Logarithmic form:
and ,
(3) Root form:
and
(4) Fraction form:
and
The shapes of the above transformations of the GPD are shown in Figure 6.
Figure 6.
Survival functions associated with transformations of the GPD .
These transformations, however, do not preserve the stability under thresholding specific to the Pareto family; as a consequence, their implementation does not lead to a stable sequential procedure. Nonetheless, they illustrate how the proposed models can be generalized in contexts where additional information on the underlying variable is available.
9. Conclusions
The splitting-based procedure presented in this article proposes an innovative experimental design for estimating an extreme failure quantile. Its development is motivated, on the one hand, by major industrial stakes, and on the other, by the limited relevance of existing methodologies in this context. The core difficulty lies in the nature of the available information: the variable of interest is latent and only exceedance indicators over chosen thresholds can be observed.
We proposed a strategy based on splitting methods to decompose the target rare event into a sequence of more tractable conditional events. The splitting formula introduces a formal decomposition that we translated into a practical sampling strategy targeting the tail of the distribution of interest progressively.
The algebraic structure of the splitting equation motivated specific parametric assumptions on the underlying distribution. We considered two models that benefit from stability properties under thresholding: a Generalized Pareto Distribution (GPD) and a Weibull distribution. Building on this structure, we designed an estimation procedure that leverages the iterative nature of the design by combining a classical maximum likelihood criterion with a backward-consistency criterion on sequentially estimated conditional quantiles. The performance of the resulting estimator was assessed numerically. Although accuracy is necessarily constrained by the quantity and quality of information, the results can still be compared—at least in terms of order of magnitude—to what could be obtained under the idealized setting of fully observed data.
From a practical standpoint, while the GPD aligns most naturally with the splitting structure (due to its exact stability under threshold conditioning), the Weibull distribution remains highly relevant for reliability modeling and engineering practice.
The applicability of the proposed framework, however, depends on the ability to implement conditional sampling schemes, which may require a preliminary analysis of the system under study. While this may limit direct applicability in some contexts, it is worth noting that methods addressing similar problems typically rely on even stronger prior knowledge of the underlying distribution. In this respect, the present approach provides a relatively generic framework, requiring more limited prior information. A concrete illustration of such an implementation is provided in [1].
Another important aspect concerns the choice of the parametric model. The assumptions introduced to ensure tractability of the sequential design are motivated by extreme value theory and are therefore primarily justified in the tail of the distribution. As a result, their adequacy may be less accurate in the earlier stages of the procedure, where the observations do not yet correspond to the extreme regime. This mismatch may induce estimation errors in the first iterations, which can propagate through the sequential updates and affect the final estimate. Although the paper outlines possible extensions toward more flexible modeling strategies, a deeper investigation of model misspecification—especially outside the extreme region—and its impact on the procedure would be a valuable direction for future work.
Beyond the specific results obtained in this work, the proposed methodology also highlights an interesting statistical perspective. The sequential nature of the procedure, together with the use of a primary estimation criterion, naturally suggests the introduction of a second statistical criterion aimed at validating the estimates at each stage.
In particular, incorporating a cross-validation-type phase within the sequential design provides an additional layer of assessment for the stability and reliability of the estimated parameters. This idea, which emerges from the structure of the method itself, opens a promising direction for future research and would require further theoretical investigation.
Author Contributions
Conceptualization, M.B. and E.M.; methodology, M.B. and E.M.; software, M.B. and E.M.; writing, M.B. and E.M. All authors have read and agreed to the published version of the manuscript.
Data Availability Statement
The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.
Conflicts of Interest
The authors declare no conflicts of interest.
Funding Statement
This research received no external funding.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Broniatowski M., Miranda E. A sequential design for extreme quantiles estimation under binary sampling. arXiv. 20192004.01563 [Google Scholar]
- 2.Weissman I. Estimation of parameters and large quantiles based on the k largest observations. J. Am. Stat. Assoc. 1978;73:812–815. [Google Scholar]
- 3.Dekkers A.L.M., Einmahl J.H.J., De Haan L. A moment estimator for the index of an extreme-value distribution. Ann. Stat. 1989;17:1833–1855. doi: 10.1214/aos/1176347397. [DOI] [Google Scholar]
- 4.De Haan L., Rootzén H. On the estimation of high quantiles. J. Stat. Plan. Inference. 1993;35:1–13. doi: 10.1016/0378-3758(93)90063-C. [DOI] [Google Scholar]
- 5.De Valk C. Approximation of high quantiles from intermediate quantiles. Extremes. 2016;19:661–684. doi: 10.1007/s10687-016-0255-3. [DOI] [Google Scholar]
- 6.De Valk C., Cai J.J. A high quantile estimator based on the log-Generalized Weibull tail limit. Econom. Stat. 2018;6:107–128. doi: 10.1016/j.ecosta.2017.03.001. [DOI] [Google Scholar]
- 7.Beirlant J., Guillou A., Dierckx G., Fils-Villetard A. Estimation of the extreme value index and extreme quantiles under random censoring. Extremes. 2007;10:151–174. doi: 10.1007/s10687-007-0039-x. [DOI] [Google Scholar]
- 8.Einmahl J.H.J., Fils-Villetard A., Guillou A. Statistics of extremes under random censoring. Bernoulli. 2008;14:207–227. doi: 10.3150/07-BEJ104. [DOI] [Google Scholar]
- 9.Worms J., Worms R. New estimators of the extreme value index under random right censoring, for heavy-tailed distributions. Extremes. 2014;17:337–358. doi: 10.1007/s10687-014-0189-6. [DOI] [Google Scholar]
- 10.Wu C.F.J. Efficient Sequential Designs with Binary Data. J. Am. Stat. Assoc. 1985;80:974–984. doi: 10.1080/01621459.1985.10478213. [DOI] [Google Scholar]
- 11.Joseph V.R. Efficient Robbins–Monro procedure for binary data. Biometrika. 2004;91:461–470. doi: 10.1093/biomet/91.2.461. [DOI] [Google Scholar]
- 12.Wu C.F.J., Tian Y. Three-phase optimal design of sensitivity experiments. J. Stat. Plan. Inference. 2014;149:1–15. doi: 10.1016/j.jspi.2013.10.007. [DOI] [Google Scholar]
- 13.Dixon W.J., Mood A.M. A Method for Obtaining and Analyzing Sensitivity Data. J. Am. Stat. Assoc. 1948;43:109–126. doi: 10.1080/01621459.1948.10483254. [DOI] [Google Scholar]
- 14.Dixon W.J. The Up-and-Down Method for Small Samples. J. Am. Stat. Assoc. 1965;60:967–978. doi: 10.1080/01621459.1965.10480843. [DOI] [Google Scholar]
- 15.O’Quigley J., Pepe M., Fisher L. Continual Reassessment Method: A Practical Design for Phase 1 Clinical Trials in Cancer. Biometrics. 1990;46:33–48. doi: 10.2307/2531628. [DOI] [PubMed] [Google Scholar]
- 16.Kahn H., Harris T.E. Estimation of Particle Transmission by Random Sampling. Natl. Bur. Stand. Appl. Math. Ser. 1951;12:27–30. [Google Scholar]
- 17.Balkema A.A., De Haan L. Residual Life Time at Great Age. Ann. Probab. 1974;2:762–804. doi: 10.1214/aop/1176996548. [DOI] [Google Scholar]
- 18.Pickands J. Statistical Inference using Extreme Order Statistics. Ann. Probab. 1975;3:119–131. [Google Scholar]
- 19.Dietrich D., De Haan L., Hüsler J. Testing extreme value conditions. Extremes. 2002;5:71–85. doi: 10.1023/A:1020934126695. [DOI] [Google Scholar]
- 20.Drees H., De Haan L., Li D. Approximations to the tail empirical distribution function with application to testing extreme value conditions. J. Stat. Plan. Inference. 2006;136:3498–3538. doi: 10.1016/j.jspi.2005.02.017. [DOI] [Google Scholar]
- 21.Choulakian V., Stephens M.A. Goodness-of-fit tests for the generalized Pareto distribution. Technometrics. 2001;43:478–484. doi: 10.1198/00401700152672573. [DOI] [Google Scholar]
- 22.Jurečková J., Picek J. A Class of Tests on the Tail Index. Extremes. 2001;4:165–183. doi: 10.1023/A:1013925226836. [DOI] [Google Scholar]
- 23.Thyrion P. Les lois exponentielles composées. Bull. Assoc. R. Actuaires Belg. 1964;62:35–44. [Google Scholar]
- 24.Thorin O. On the infinite divisibility of the Pareto distribution. Scand. Actuar. J. 1977;1:31–40. doi: 10.1080/03461238.1977.10405623. [DOI] [Google Scholar]
- 25.Feller W. An Introduction to Probability Theory and Its Applications. Volume 2 Wiley; Hoboken, NJ, USA: 1971. [Google Scholar]
- 26.Naveau P., Huser R., Ribereau P., Hannart A. modelling jointly low, moderate, and heavy rainfall intensities without a threshold selection. Water Resour. Res. 2016;52:2753–2769. doi: 10.1002/2015WR018552. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.






