The starting point of the USP 1033 guideline is the requirement to measure product potency within its specification limits during routine testing. Meanwhile, the pharmaceutical industry traditionally defines acceptance criteria for the measuring analytical procedure in terms of either accuracy and precision or total analytical error (TAE) and risk.
Every analyst and even USP 1033 authors struggle with reconciling these two concepts because it is not clear how to translate product requirements to procedure requirements. Latest (30-SEP-2024) USP 1033 draft addresses this by making a simplifying assumption: that the production process exhibits no variability, allowing product specifications to be directly expressed through TAE. In practice analysts then rely on these results (i.e. formulas) and add rule-of-thumb margins (e.g. Six Sigma) to account for actual process variability. However such approaches, often lacking a theoretical foundation, can break down in edge cases or lead to overly strict acceptance criteria.
All this raises an important question: can procedure acceptance criteria be correctly derived from product specifications? The answer is yes, but not as in the traditional sense as global scalar limits. Based on some recent work and in line with USP 1033, I’ll explain here briefly the exact (!) method of assessing procedure performance based on product specifications. An example application is also available here.
The first step is spelling out the assumptions and abstracting them to a formal framework. From USP 1033 concepts and formulas we can deduce the following measurement model (and vice versa):
where is the geometric center of the production process,
the recovery of the procedure, and
,
and
the unit-centered lognormal random variables of respectively the production process variability (
), and the between and within run variability of the procedure (
). Note that production process parameters (i.e.
and
) are assumed known and often set to some safe estimate (cf. USP 1033). Also note that procedure parameters
,
and
are determined during validation at different levels of the true value and thus are a function of it. True value during routine measurement is
.
The second step is stating the problem to be solved which translates to:
or in layman’s terms: The probability to measure outside of product upper (USL) and lower (LSL) specification limit should not exceed . To make it more concrete, we can translate this into a comprehensible analytical target profile (ATP) which I base here on the latest (30-SEP-2024) USP 1033 draft example:
The procedure must be able to quantify relative potency (RP) in a range from 0.5 RP to 2 RP such that, under the assumed production process geometrically centered on 1 RP with 4.77 % SD, one can expect less than 1 % of all future measured values to fall outside the [0.70; 1.43] RP product specification range.
The third step is then deriving from the second equation the acceptance limits for the procedure. This is where it becomes clear that there is no unique global solution for accuracy and precision, or TAE and risk due to the interactions between and
and their dependence on the true value. (The suggested solutions in USP 1033 are ones-of-many and because of that may result in falsely rejecting a perfectly fine procedure.) More complex criteria are required. In the graph that follows the validation of the procedure is represented as a good compromise between complexity and intelligibility.

Figure 1: The experimental measurements (grey circles) are plotted as relative bias (RB) in function of relative potency (RP). The red dotted curve is the expected relative bias and the blue dotted interval is the added intermediate precision (IP). (These estimates are taken from the USP 1033 tables.) One can state roughly that the blue dotted interval covers about 68 % of the measurements. The density curves below reflect the performance in routine. The green density represents the true RP of the production process. The blue density tells us what will happen when we measure these products with our procedure (i.e. based on the procedure’s performance expressed with the blue dotted lines). One can see that the measurements would remain well within the boundaries of the product specification (i.e. the yellow bars). The black density has 99 % of its area within the yellow bars, which then translates to the black dotted acceptance interval of the procedure as the maximal addition of global IP while still meeting the requirements. Hence the difference between black and blue dotted lines can be interpreted as maximum global IP that the method can incur while still remaining within ATP specification.
The above graph represents an exact solution based on the assumptions in step 1 and step 2, which among others implies:
– that the procedure performance’s dependence on true value is taken into consideration (hence the importance of interpolation of procedure performance estimates over the whole working range),
– that the procedure performance is “weighted” according to the production process density,
– that lognormality (although hardly visible) is taken into account, etc.
The graph also can be made interactive (cf. here) so the analyst can adjust various components such as the production process, procedure performance characteristics (locally or globally), confidence levels of the tolerance intervals, etc. and gain immediate feedback on its effect in routine use. This allows the analyst to find the best way (in terms of effort versus impact) to make the procedure fit for its purpose.
PS And yes, all this can also be applied to ICH Q2(R2) by using a measurement model that is based on the normal distribution.