D- and A-Optimal Screening Designs Jonathan Stallrich North Carolina State University Department of Statistics

2025-05-06 0 0 2.04MB 38 页 10玖币
侵权投诉
D- and A-Optimal Screening Designs
Jonathan Stallrich
North Carolina State University, Department of Statistics
Katherine Allen-Moyer
North Carolina State University, Department of Statistics
Bradley Jones
JMP Statistical Discovery Software LLC
November 1, 2022
Abstract
In practice, optimal screening designs for arbitrary run sizes are traditionally generated
using the D-criterion with factor settings fixed at ±1, even when considering continu-
ous factors with levels in [1,1]. This paper identifies cases of undesirable estimation
variance properties for such D-optimal designs and argues that generally A-optimal de-
signs tend to push variances closer to their minimum possible value. New insights about
the behavior of the criteria are found through a study of their respective coordinate-
exchange formulas. The study confirms the existence of D-optimal designs comprised
only of settings ±1 for both main effect and interaction models for blocked and un-
blocked experiments. Scenarios are also identified for which arbitrary manipulation of
a coordinate between [1,1] leads to infinitely many D-optimal designs each having
different variance properties. For the same conditions, the A-criterion is shown to have
a unique optimal coordinate value for improvement. We also compare Bayesian version
of the A- and D-criteria in how they balance minimization of estimation variance and
bias. Multiple examples of screening designs are considered for various models under
Bayesian and non-Bayesian versions of the A- and D-criteria.
Keywords: Bayesian optimal design; blocking; continuous exchange algorithm; D-optimality;
factorial experiments; minimum aliasing
1
arXiv:2210.13943v2 [stat.ME] 31 Oct 2022
1 Introduction
A screening experiment is an initial step in a sequential experimental procedure to un-
derstand and/or optimize a process dependent upon many controllable factors. Such ex-
periments are common in pharmaceuticals, agriculture, genetics, defense, and textiles (see
Dean and Lewis (2006) for a comprehensive overview of screening design methodology and
applications). The screening analysis aims to identify the few factors that drive most of
the process variation often according to a linear model comprised of main effects, interac-
tion effects, and, in the case of numeric factors, quadratic effects (Jones and Nachtsheim,
2011a). Each effect corresponds to one or more factors, and a factor is said to be active if
at least one of its corresponding effects is large relative to the process noise; otherwise the
factor is said to be inert. Analyses under this class of models often follow effect principles
of sparsity, hierarchy, and heredity (see Chapter 9 of Wu and Hamada (2009)), with the
primary goal of correctly classifying each factor as active or inert.
A screening design is represented by an n×kmatrix, Xd, with rows xT
i= (xi1, . . . , xik)
where xij represents the j-th factor’s setting for run i. To standardize screening designs
across applications, continuous factor settings are scaled so xij [1,1] while categorical
factor settings are often restricted to two levels, making xij =±1. We compare Xd’s based
on the statistical properties of the effects’ least-squares estimators because their properties
are tractable, particularly their variances and potential biases. The goal then is to identify
an Xdthat minimizes the individual variances and biases of these effect estimators.
Suppose the model is correctly specified and there are designs having unique least-
squares estimators for all effects. Then these estimators are unbiased and designs may
be compared based on their estimation variances. A design having variances that are as
small as possible will improve one’s ability to correctly classify factors as active/inert. For
models comprised solely of main effects and interactions, orthogonal designs have estimation
variances simultaneously equal to their minimum possible value across all designs. Such
designs exist only when nis a multiple of 4; for other nit is unclear which design will
have the best variance properties. Still, designs should be compared based on how close
their variances are to their respective minimum possible values. This approach requires
knowledge of the minimum values as well as some measure of closeness.
One approach for identifying minimum variances is to approximate them using the
theoretical value assuming an orthogonal design exists, but such values may be unattainable.
2
The c-criterion (Atkinson et al., 2007) may be used to identify the minimum variance for a
given effect, but without any guarantee of the estimability of the other effects of interest. To
remedy this estimability issue, Allen-Moyer and Stallrich (2022) proposed the cE-criterion
to calculate these minimum variances exactly. It is less clear how to measure the proximity
of a design’s variances to their cEvalues. The Pareto frontier approach by Lu et al. (2011) is
well-suited for this problem but can be cumbersome in practice. A more practical solution
is to evaluate and rank designs according to a single criterion that involves a scalar measure
of all the variances. Such criteria should be straightforward to evaluate and optimize, and
the resulting optimal designs should have variances close to their cEvalues. Different forms
of the D- and A-criterion (see Section 2.1) are popular variance-based criteria employed in
the screening design literature and will be the focus of this paper.
Designs that optimize D- and A-criteria can coincide for some n, but this does not
mean the criteria equivalently summarize variances. Consider a screening problem with
n= 7 runs and k= 5 factors that assumes a main effect model. It is well-known that there
always exists a D-optimal design comprised of xij =±1, even when xij [1,1] (Box and
Draper, 1971). While other D-optimal designs having xij (1,1) may exist, the screening
literature predominantly fixes xij =±1 with no assumed degradation to the resulting
variances. For example, Jones et al. (2020a) found an A-optimal design with xij values of
±1 and 0 having smaller variances compared to D-optimal designs comprised of xij =±1
only. Figure 1 shows this A-optimal design, which has x14 =x15 = 0. Figure 1 also shows
the corresponding main effect variances (in ascending order) of the A-optimal design and
two D-optimal designs comprised of xij =±1. The minimum possible variances assuming
an orthogonal design exists are 1/7=0.1429 and the minimum variances under the cE-
criterion from Allen-Moyer and Stallrich (2022) are 0.1459. Each of the A-optimal design’s
variances are equal to or smaller than the two competing D-optimal designs comprised of
±1.
As it turns out, the A-optimal design in Figure 1 is also D-optimal despite having some
xij = 0. In fact, changing either x14 or x15 to any value in [1,1] produces yet another
D-optimal design but with equal or larger variances than the A-optimal design. The A-
optimal design in this case, however, is unique. The existence of infinitely many D-optimal
designs, each with equal or larger variances relative to the A-optimal design, is cause for
concern about utilizing the D-criterion to rank screening designs. In this example, the
3
11100
-1 -1 1 -1 1
-1 1 -1 -1 1
1 -1 -1 -1 -1
-1 -1 1 1 -1
1 -1 -1 1 1
-1 1 -1 1 -1
Figure 1: (Left) n= 7, k = 5, A-optimal design. (Right) Main effect variances (in ascending
order) for A- and D-optimal designs. The design “D-optimal 1,1” replaces x14 and x15 of
left design with 1. Design “D-optimal 1,1” is similarly defined. The minimum possible
variances assuming an orthogonal design would each be 1/7=0.1429 and the minimum
variances under the cE-criterion from Allen-Moyer and Stallrich (2022) are 0.1459.
A-criterion was better able to differentiate designs in terms of their ability to minimize the
main effect variances simultaneously.
This is not to say D-optimal designs are less valuable than A-optimal designs. Such
designs have been used with great success in practice and the relative differences of the
variances in Figure 1 do not appear large. Whether these differences impact the analysis
depends on the ratio of the true main effect, denoted βj, and the process variance, σ2. When
performing a two-sided t-test for the null hypothesis βj= 0, the associated noncentrality
parameter will be βjdivided by the square root of the variances shown in Figure 1.
When βjis large, slight differences in the variances will not affect the noncentrality
parameter, and hence will not affect power of the tests. The differences in variances will
have a significant impact as βjgets smaller. For example, suppose βj= 1 and we
perform a t-test for β1= 0 with significance level α= 0.05. The power for this test under
the D-optimal design with x14 =x15 = 1 is 0.6355 while for the A-optimal design it is
0.7135. Without any prior knowledge of the βj, it is important then to find a design that
decreases the individual variances as much as possible.
Based on the effect principles, it is common to fit a main effect model even though
interactions and/or quadratic effects may be active. The least-squares estimators for the
main effect model may then become biased. Rather than try to estimate all potentially
important effects, one can quantify the bias of the estimators and identify a design that
4
simultaneously reduces estimation variance and bias. Let βbe the vector of the largest
collection of effects that may be important and hence captures the true model. Partition
βinto β1and β2where β1are effects we believe are most likely to be important and
correspond to the effects in the fitted model, and β2are the remaining effects that are
potentially important but ignored in the fitted model. The possible bias from estimating
β1under the fitted model when the true model includes all βis 2where Ais the
design’s so-called alias matrix. DuMouchel and Jones (1994) construct designs under model
uncertainty by assigning a prior distribution to β1and β2, and ranking designs according
to the D-criterion applied to β’s posterior covariance matrix. While Bayesian D-optimal
designs have shown an ability to balance minimizing bias and variance, the possible flaws
of the D-criterion pointed out earlier are still concerning. Better designs may then be
found with a Bayesian A-criterion, which has not received much attention in the screening
literature.
This paper makes two important contributions that build a strong case for construct-
ing screening designs under different forms of the A-criterion. The first contribution is a
comparison of the behavior of the D- and A-criteria in response to manipulating a single
coordinate of a given design. Our investigation not only provides insights into the criteria’s
coordinate exchange algorithms, a popular design construction algorithm, but also estab-
lishes the existence of D-optimal designs with xij =±1 for models including main effects
and/or interactions, as well as nuisance effects, such as block effects. We are only aware of
such a result for main effect models with an intercept. We also identify cases in which the
D-criterion is invariant to any possible coordinate exchange, meaning the D-criterion con-
siders all such designs as having equal value despite potentially having different variances.
For such cases, we show that the A-criterion has a unique optimal coordinate exchange. Our
second contribution is the promotion of a weighted Bayesian A-criterion for constructing
designs that balance bias and variance minimization. We compare new screening designs
generated under coordinate-exchange algorithms for common factorial models and show
the Bayesian A-optimal designs have more appealing variance and bias properties than
Bayesian D-optimal designs.
The paper is organized as follows. Section 2 reviews traditional and current screening
models and criteria. Section 3 investigates the behavior of the D- and A-criteria following
coordinate exchanges to an existing design for models including nuisance effects. It also
5
摘要:

D-andA-OptimalScreeningDesignsJonathanStallrichNorthCarolinaStateUniversity,DepartmentofStatisticsKatherineAllen-MoyerNorthCarolinaStateUniversity,DepartmentofStatisticsBradleyJonesJMPStatisticalDiscoverySoftwareLLCNovember1,2022AbstractInpractice,optimalscreeningdesignsforarbitraryrunsizesaretradit...

展开>> 收起<<
D- and A-Optimal Screening Designs Jonathan Stallrich North Carolina State University Department of Statistics.pdf

共38页,预览5页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!
分类:图书资源 价格:10玖币 属性:38 页 大小:2.04MB 格式:PDF 时间:2025-05-06

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 38
客服
关注