Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Optimal experimental design

From Wikipedia, the free encyclopedia
(Redirected fromOptimal design)
Experimental design that is optimal with respect to some statistical criterion
This article is about optimaldesign of experiments. For optimal design in control theory, seeshape optimization.
Picture of a man taking measurements with a theodolite in a frozen environment.
Gustav Elfving developed the optimal design of experiments, and so minimized surveyors' need fortheodolite measurements(pictured), while trapped in his tent in storm-riddenGreenland.[1]

In thedesign of experiments,optimal experimental designs (oroptimum designs[2]) are a class ofexperimental designs that areoptimal with respect to somestatisticalcriterion. The creation of this field of statistics has been credited to Danish statisticianKirstine Smith.[3][4]

In thedesign of experiments forestimatingstatistical models,optimal designs allow parameters to beestimated without bias and withminimum variance. A non-optimal design requires a greater number ofexperimental runs toestimate theparameters with the sameprecision as an optimal design. In practical terms, optimal experiments can reduce the costs of experimentation.

The optimality of a design depends on thestatistical model and is assessed with respect to a statistical criterion, which is related to the variance-matrix of the estimator. Specifying an appropriate model and specifying a suitable criterion function both require understanding ofstatistical theory and practical knowledge withdesigning experiments.

Advantages

[edit]

Optimal designs offer three advantages over sub-optimalexperimental designs:[5]

  1. Optimal designs reduce the costs of experimentation by allowingstatistical models to be estimated with fewer experimental runs.
  2. Optimal designs can accommodate multiple types of factors, such as process, mixture, and discrete factors.
  3. Designs can be optimized when the design-space is constrained, for example, when the mathematical process-space contains factor-settings that are practically infeasible (e.g. due to safety concerns).

Minimizing the variance of estimators

[edit]

Experimental designs are evaluated using statistical criteria.[6]

It is known that theleast squares estimator minimizes thevariance ofmean-unbiasedestimators (under the conditions of theGauss–Markov theorem). In theestimation theory forstatistical models with onerealparameter, thereciprocal of the variance of an ("efficient") estimator is called the "Fisher information" for that estimator.[7] Because of this reciprocity,minimizing thevariance corresponds tomaximizing theinformation.

When thestatistical model has severalparameters, however, themean of the parameter-estimator is avector and itsvariance is amatrix. Theinverse matrix of the variance-matrix is called the "information matrix". Because the variance of the estimator of a parameter vector is a matrix, the problem of "minimizing the variance" is complicated. Usingstatistical theory, statisticians compress the information-matrix using real-valuedsummary statistics; being real-valued functions, these "information criteria" can be maximized.[8] The traditional optimality-criteria areinvariants of theinformation matrix; algebraically, the traditional optimality-criteria arefunctionals of theeigenvalues of the information matrix.

  • A-optimality ("average" ortrace)
    • One criterion isA-optimality, which seeks to minimize thetrace of theinverse of the information matrix. This criterion results in minimizing the average variance of the estimates of the regression coefficients.
  • C-optimality
  • D-optimality (determinant)
  • E-optimality (eigenvalue)
    • Another design isE-optimality, which maximizes the minimumeigenvalue of the information matrix.
  • S-optimality[9]
    • This criterion maximizes a quantity measuring the mutual column orthogonality of X and thedeterminant of the information matrix.
  • T-optimality
    • This criterion maximizes the discrepancy between two proposed models at the design locations.[10]

Other optimality-criteria are concerned with the variance ofpredictions:

  • G-optimality
    • A popular criterion isG-optimality, which seeks to minimize the maximum entry in thediagonal of thehat matrix X(X'X)−1X'. This has the effect of minimizing the maximum variance of the predicted values.
  • I-optimality (integrated)
    • A second criterion on prediction variance isI-optimality, which seeks to minimize the average prediction varianceover the design space.
  • V-optimality (variance)
    • A third criterion on prediction variance isV-optimality, which seeks to minimize the average prediction variance over a set of m specific points.[11]

Contrasts

[edit]
Main article:Contrast (statistics)
See also:Nuisance parameter

In many applications, the statistician is most concerned with a"parameter of interest" rather than with"nuisance parameters". More generally, statisticians considerlinear combinations of parameters, which are estimated via linear combinations of treatment-means in thedesign of experiments and in theanalysis of variance; such linear combinations are calledcontrasts. Statisticians can use appropriate optimality-criteria for suchparameters of interest and forcontrasts.[12]

Implementation

[edit]

Catalogs of optimal designs occur in books and in software libraries.

In addition, majorstatistical systems likeSAS andR have procedures for optimizing a design according to a user's specification. The experimenter must specify amodel for the design and an optimality-criterion before the method can compute an optimal design.[13]

Practical considerations

[edit]

Some advanced topics in optimal design require morestatistical theory and practical knowledge in designing experiments.

Model dependence and robustness

[edit]

Since the optimality criterion of most optimal designs is based on some function of the information matrix, the 'optimality' of a given design ismodel dependent: While an optimal design is best for thatmodel, its performance may deteriorate on othermodels. On othermodels, anoptimal design can be either better or worse than a non-optimal design.[14] Therefore, it is important tobenchmark the performance of designs under alternativemodels.[15]

Choosing an optimality criterion and robustness

[edit]

The choice of an appropriate optimality criterion requires some thought, and it is useful to benchmark the performance of designs with respect to several optimality criteria. Cornell writes that

since the [traditional optimality] criteria . . . are variance-minimizing criteria, . . . a design that is optimal for a given model using one of the . . . criteria is usually near-optimal for the same model with respect to the other criteria.

— [16]

Indeed, there are several classes of designs for which all the traditional optimality-criteria agree, according to the theory of "universal optimality" ofKiefer.[17] The experience of practitioners like Cornell and the "universal optimality" theory of Kiefer suggest that robustness with respect to changes in theoptimality-criterion is much greater than is robustness with respect to changes in themodel.

Flexible optimality criteria and convex analysis

[edit]

High-quality statistical software provide a combination of libraries of optimal designs or iterative methods for constructing approximately optimal designs, depending on the model specified and the optimality criterion. Users may use a standard optimality-criterion or may program a custom-made criterion.

All of the traditional optimality-criteria areconvex (or concave) functions, and therefore optimal-designs are amenable to the mathematical theory ofconvex analysis and their computation can use specialized methods ofconvex minimization.[18] The practitioner need not selectexactly one traditional, optimality-criterion, but can specify a custom criterion. In particular, the practitioner can specify a convex criterion using the maxima of convex optimality-criteria andnonnegative combinations of optimality criteria (since these operations preserveconvex functions). Forconvex optimality criteria, theKiefer-Wolfowitzequivalence theorem allows the practitioner to verify that a given design is globally optimal.[19] TheKiefer-Wolfowitzequivalence theorem is related with theLegendre-Fenchelconjugacy forconvex functions.[20]

If an optimality-criterion lacksconvexity, then finding aglobal optimum and verifying its optimality often are difficult.

Model uncertainty and Bayesian approaches

[edit]

Model selection

[edit]
See also:Model selection

When scientists wish to test several theories, then a statistician can design an experiment that allows optimal tests between specified models. Such "discrimination experiments" are especially important in thebiostatistics supportingpharmacokinetics andpharmacodynamics, following the work ofCox and Atkinson.[21]

Bayesian experimental design

[edit]
Main article:Bayesian experimental design

When practitioners need to consider multiplemodels, they can specify aprobability-measure on the models and then select any design maximizing theexpected value of such an experiment. Such probability-based optimal-designs are called optimalBayesiandesigns. SuchBayesian designs are used especially forgeneralized linear models (where the response follows anexponential-family distribution).[22]

The use of aBayesian design does not force statisticians to useBayesian methods to analyze the data, however. Indeed, the "Bayesian" label for probability-based experimental-designs is disliked by some researchers.[23] Alternative terminology for "Bayesian" optimality includes "on-average" optimality or "population" optimality.

Iterative experimentation

[edit]

Scientific experimentation is an iterative process, and statisticians have developed several approaches to the optimal design of sequential experiments.

Sequential analysis

[edit]
Main article:Sequential analysis

Sequential analysis was pioneered byAbraham Wald.[24] In 1972,Herman Chernoff wrote an overview of optimal sequential designs,[25] whileadaptive designs were surveyed later by S. Zacks.[26] Of course, much work on the optimal design of experiments is related to the theory ofoptimal decisions, especially thestatistical decision theory ofAbraham Wald.[27]

Response-surface methodology

[edit]
Main article:Response surface methodology

Optimal designs forresponse-surface models are discussed in the textbook by Atkinson, Donev and Tobias, and in the survey of Gaffke and Heiligers and in the mathematical text of Pukelsheim. Theblocking of optimal designs is discussed in the textbook of Atkinson, Donev and Tobias and also in the monograph by Goos.

The earliest optimal designs were developed to estimate the parameters of regression models with continuous variables, for example, byJ. D. Gergonne in 1815 (Stigler). In English, two early contributions were made byCharles S. Peirce andKirstine Smith.

Pioneering designs for multivariateresponse-surfaces were proposed byGeorge E. P. Box. However, Box's designs have few optimality properties. Indeed, theBox–Behnken design requires excessive experimental runs when the number of variables exceeds three.[28]Box's"central-composite" designs require more experimental runs than do the optimal designs of Kôno.[29]

System identification and stochastic approximation

[edit]
See also:System identification andStochastic approximation

The optimization of sequential experimentation is studied also instochastic programming and insystems andcontrol. Popular methods includestochastic approximation and other methods ofstochastic optimization. Much of this research has been associated with the subdiscipline ofsystem identification.[30]In computationaloptimal control, D. Judin & A. Nemirovskii andBoris Polyak has described methods that are more efficient than the (Armijo-style)step-size rules introduced byG. E. P. Box inresponse-surface methodology.[31]

Adaptive designs are used inclinical trials, and optimaladaptive designs are surveyed in theHandbook of Experimental Designs chapter by Shelemyahu Zacks.

Specifying the number of experimental runs

[edit]

Using a computer to find a good design

[edit]

There are several methods of finding an optimal design, given ana priori restriction on the number of experimental runs or replications. Some of these methods are discussed by Atkinson, Donev and Tobias and in the paper by Hardin andSloane. Of course, fixing the number of experimental runsa priori would be impractical. Prudent statisticians examine the other optimal designs, whose number of experimental runs differ.

Discretizing probability-measure designs

[edit]

In the mathematical theory on optimal experiments, an optimal design can be aprobability measure that issupported on an infinite set of observation-locations. Such optimal probability-measure designs solve a mathematical problem that neglected to specify the cost of observations and experimental runs. Nonetheless, such optimal probability-measure designs can bediscretized to furnishapproximately optimal designs.[32]

In some cases, a finite set of observation-locations suffices tosupport an optimal design. Such a result was proved by Kôno andKiefer in their works onresponse-surface designs for quadratic models. The Kôno–Kiefer analysis explains why optimal designs for response-surfaces can have discrete supports, which are very similar as do the less efficient designs that have been traditional inresponse surface methodology.[33]

History

[edit]

In 1815, an article on optimal designs forpolynomial regression was published byJoseph Diaz Gergonne, according toStigler.

Charles S. Peirce proposed an economic theory of scientific experimentation in 1876, which sought to maximize the precision of the estimates. Peirce's optimal allocation immediately improved the accuracy of gravitational experiments and was used for decades by Peirce and his colleagues. In his 1882 published lecture atJohns Hopkins University, Peirce introduced experimental design with these words:

Logic will not undertake to inform you what kind of experiments you ought to make in order best to determine the acceleration of gravity, or the value of the Ohm; but it will tell you how to proceed to form a plan of experimentation.

[....] Unfortunately practice generally precedes theory, and it is the usual fate of mankind to get things done in some boggling way first, and find out afterward how they could have been done much more easily and perfectly.[34]

Kirstine Smith proposed optimal designs for polynomial models in 1918. (Kirstine Smith had been a student of the Danish statisticianThorvald N. Thiele and was working withKarl Pearson in London.)

See also

[edit]

Notes

[edit]
  1. ^Nordström (1999, p. 176)
  2. ^The adjective "optimum" (and not "optimal") "is the slightly older form in English and avoids the construction 'optim(um) + al´—there is no 'optimalis' in Latin" (page x inOptimum Experimental Designs, with SAS, by Atkinson, Donev, and Tobias).
  3. ^Guttorp, P.; Lindgren, G. (2009). "Karl Pearson and the Scandinavian school of statistics".International Statistical Review.77: 64.CiteSeerX 10.1.1.368.8328.doi:10.1111/j.1751-5823.2009.00069.x.S2CID 121294724.
  4. ^Smith, Kirstine (1918)."On the standard deviations of adjusted and interpolated values of an observed polynomial function and its constants and the guidance they give towards a proper choice of the distribution of observations".Biometrika.12 (1/2):1–85.doi:10.2307/2331929.JSTOR 2331929.
  5. ^These three advantages (of optimal designs) are documented in the textbook by Atkinson, Donev, and Tobias.
  6. ^Such criteria are calledobjective functions inoptimization theory.
  7. ^TheFisher information and other "information"functionals are fundamental concepts instatistical theory.
  8. ^Traditionally, statisticians have evaluated estimators and designs by considering somesummary statistic of the covariance matrix (of amean-unbiased estimator), usually with positive real values (like thedeterminant ormatrix trace). Working with positive real-numbers brings several advantages: If the estimator of a single parameter has a positive variance, then the variance and the Fisher information are both positive real numbers; hence they are members of the convex cone of nonnegative real numbers (whose nonzero members have reciprocals in this same cone).
    For several parameters, the covariance-matrices and information-matrices are elements of the convex cone of nonnegative-definite symmetric matrices in apartiallyordered vector space, under theLoewner (Löwner) order. This cone is closed under matrix-matrix addition, under matrix-inversion, and under the multiplication of positive real-numbers and matrices.An exposition of matrix theory and the Loewner-order appears in Pukelsheim.
  9. ^Shin, Yeonjong; Xiu, Dongbin (2016). "Nonadaptive quasi-optimal points selection for least squares linear regression".SIAM Journal on Scientific Computing.38 (1):A385–A411.Bibcode:2016SJSC...38A.385S.doi:10.1137/15M1015868.
  10. ^Atkinson, A. C.; Fedorov, V. V. (1975)."The design of experiments for discriminating between two rival models".Biometrika.62 (1):57–70.doi:10.1093/biomet/62.1.57.ISSN 0006-3444.
  11. ^The above optimality-criteria are convex functions on domains ofsymmetric positive-semidefinite matrices: See an on-line textbook for practitioners, which has many illustrations and statistical applications:Boyd and Vandenberghe discuss optimal experimental designs on pages 384–396.
  12. ^Optimality criteria for"parameters of interest" and forcontrasts are discussed by Atkinson, Donev and Tobias.
  13. ^Iterative methods and approximation algorithms are surveyed in the textbook by Atkinson, Donev and Tobias and in the monographs of Fedorov (historical) and Pukelsheim, and in the survey article by Gaffke and Heiligers.
  14. ^See Kiefer ("Optimum Designs for Fitting Biased Multiresponse Surfaces" pages 289–299).
  15. ^Such benchmarking is discussed in the textbook by Atkinson et al. and in the papers of Kiefer.Model-robust designs (including "Bayesian" designs) are surveyed by Chang and Notz.
  16. ^Cornell, John (2002).Experiments with Mixtures: Designs, Models, and the Analysis of Mixture Data (third ed.). Wiley.ISBN 978-0-471-07916-3. (Pages 400-401)
  17. ^An introduction to "universal optimality" appears in the textbook of Atkinson, Donev, and Tobias. More detailed expositions occur in the advanced textbook of Pukelsheim and the papers of Kiefer.
  18. ^Computational methods are discussed by Pukelsheim and by Gaffke and Heiligers.
  19. ^TheKiefer-Wolfowitzequivalence theorem is discussed in Chapter 9 of Atkinson, Donev, and Tobias.
  20. ^Pukelsheim usesconvex analysis to studyKiefer-Wolfowitzequivalence theorem in relation to theLegendre-Fenchelconjugacy forconvex functionsTheminimization ofconvex functions on domains ofsymmetric positive-semidefinite matrices is explained in an on-line textbook for practitioners, which has many illustrations and statistical applications:Boyd and Vandenberghe discuss optimal experimental designs on pages 384–396.
  21. ^See Chapter 20 in Atkinison, Donev, and Tobias.
  22. ^Bayesian designs are discussed in Chapter 18 of the textbook by Atkinson, Donev, and Tobias. More advanced discussions occur in the monograph by Fedorov and Hackl, and the articles by Chaloner and Verdinelli and by DasGupta.Bayesian designs and other aspects of "model-robust" designs are discussed by Chang and Notz.
  23. ^As an alternative to "Bayesian optimality", "on-average optimality" is advocated in Fedorov and Hackl.
  24. ^Wald, Abraham (June 1945)."Sequential Tests of Statistical Hypotheses".The Annals of Mathematical Statistics.16 (2):117–186.doi:10.1214/aoms/1177731118.JSTOR 2235829.
  25. ^Chernoff, H. (1972)Sequential Analysis and Optimal Design, SIAM Monograph.
  26. ^Zacks, S. (1996) "Adaptive Designs for Parametric Models". In: Ghosh, S. and Rao, C. R., (Eds) (1996).Design and Analysis of Experiments, Handbook of Statistics, Volume 13. North-Holland.ISBN 0-444-82061-2. (pages 151–180)
  27. ^Henry P. Wynn wrote, "the modern theory of optimum design has its roots in the decision theory school of U.S. statistics founded byAbraham Wald" in his introduction "Jack Kiefer's Contributions to Experimental Design", which is pages xvii–xxiv in the following volume:Kiefer acknowledges Wald's influence and results on many pages – 273 (page 55 in the reprinted volume), 280 (62), 289-291 (71-73), 294 (76), 297 (79), 315 (97) 319 (101) – in this article:
  28. ^In the field ofresponse surface methodology, theinefficiency of theBox–Behnken design is noted by Wu and Hamada (page 422).
    • Wu, C. F. Jeff & Hamada, Michael (2002).Experiments: Planning, Analysis, and Parameter Design Optimization. Wiley.ISBN 978-0-471-25511-6.
    Optimal designs for "follow-up" experiments are discussed by Wu and Hamada.
  29. ^Theinefficiency ofBox's"central-composite" designs are discussed by according to Atkinson, Donev, and Tobias (page 165). These authors also discuss theblocking of Kôno-type designs for quadraticresponse-surfaces.
  30. ^In system identification, the following books have chapters on optimal experimental design:
  31. ^Some step-size rules for of Judin & Nemirovskii and ofPolyakArchived 2007-10-31 at theWayback Machine are explained in the textbook by Kushner and Yin:
  32. ^Thediscretization of optimal probability-measure designs to provideapproximately optimal designs is discussed by Atkinson, Donev, and Tobias and by Pukelsheim (especially Chapter 12).
  33. ^Regarding designs for quadraticresponse-surfaces, the results of Kôno andKiefer are discussed in Atkinson, Donev, and Tobias.Mathematically, such results are associated withChebyshev polynomials, "Markov systems", and "moment spaces": See
  34. ^Peirce, C. S. (1882), "Introductory Lecture on the Study of Logic" delivered September 1882, published inJohns Hopkins University Circulars, v. 2, n. 19, pp. 11–12, November 1882, see p. 11,Google BooksEprint. Reprinted inCollected Papers v. 7, paragraphs 59–76, see 59, 63,Writings of Charles S. Peirce v. 4, pp. 378–82, see 378, 379, andThe Essential Peirce v. 1, pp. 210–14, see 210–1, also lower down on 211.

References

[edit]

Further reading

[edit]

Textbooks for practitioners and students

[edit]

Textbooks emphasizing regression and response-surface methodology

[edit]

The textbook by Atkinson, Donev and Tobias has been used for short courses for industrial practitioners as well as university courses.

Textbooks emphasizing block designs

[edit]

Optimalblock designs are discussed by Bailey and by Bapat. The first chapter of Bapat's book reviews thelinear algebra used by Bailey (or the advanced books below). Bailey's exercises and discussion ofrandomization both emphasize statistical concepts (rather than algebraic computations).

Optimalblock designs are discussed in the advanced monograph by Shah and Sinha and in the survey-articles by Cheng and by Majumdar.

Books for professional statisticians and researchers

[edit]

Articles and chapters

[edit]

Historical

[edit]
Scientific
method
Treatment
andblocking
Models
andinference
Designs

Completely
randomized
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Optimal_experimental_design&oldid=1306689944"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp