Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Grey box model

From Wikipedia, the free encyclopedia
Mathematical data production model with limited structure

Inmathematics,statistics, andcomputational modelling, agrey box model[1][2][3][4] combines a partial theoretical structure with data to complete the model. The theoretical structure may vary from information on the smoothness of results, to models that need only parameter values from data or existing literature.[5] Thus, almost all models are grey box models as opposed toblack box where no model form is assumed orwhite box models that are purely theoretical. Some models assume a special form such as alinear regression[6][7] orneural network.[8][9] These have special analysis methods. In particularlinear regression techniques[10] are much more efficient than most non-linear techniques.[11][12] The model can bedeterministic orstochastic (i.e. containing random components) depending on its planned use.

Model form

[edit]

The general case is anon-linear model with a partial theoretical structure and some unknown parts derived from data. Models with unlike theoretical structures need to be evaluated individually,[1][13][14] possibly usingsimulated annealing orgenetic algorithms.

Within a particular model structure,parameters[14][15] or variable parameter relations[5][16] may need to be found. For a particular structure it is arbitrarily assumed that the data consists of sets of feed vectorsf, product vectorsp, and operating condition vectorsc.[5] Typicallyc will contain values extracted fromf, as well as other values. In many cases a model can be converted to a function of the form:[5][17][18]

m(f,p,q)

where the vector functionm gives the errors between the datap, and the model predictions. The vectorq gives some variable parameters that are the model's unknown parts.

The parametersq vary with the operating conditionsc in a manner to be determined.[5][17] This relation can be specified asq =Ac whereA is a matrix of unknown coefficients, andc as inlinear regression[6][7] includes a constant term and possibly transformed values of the original operating conditions to obtain non-linear relations[19][20] between the original operating conditions andq. It is then a matter of selecting which terms inA are non-zero and assigning their values. The model completion becomes anoptimization problem to determine the non-zero values inA that minimizes the error termsm(f,p,Ac) over the data.[1][16][21][22][23]

Model completion

[edit]

Once a selection of non-zero values is made, the remaining coefficients inA can be determined by minimizingm(f,p,Ac) over the data with respect to the nonzero values inA, typically bynon-linear least squares. Selection of the nonzero terms can be done by optimization methods such assimulated annealing andevolutionary algorithms. Also thenon-linear least squares can provide accuracy estimates[11][15] for the elements ofA that can be used to determine if they are significantly different from zero, thus providing a method ofterm selection.[24][25]

It is sometimes possible to calculate values ofq for each data set, directly or bynon-linear least squares. Then the more efficientlinear regression can be used to predictq usingc thus selecting the non-zero values inA and estimating their values. Once the non-zero values are locatednon-linear least squares can be used on the original modelm(f,p,Ac) to refine these values .[16][21][22]

A third method ismodel inversion,[5][17][18] which converts the non-linearm(f,p,Ac) into an approximate linear form in the elements ofA, that can be examined using efficient term selection[24][25] and evaluation of the linear regression.[10] For the simple case of a singleq value (q =aTc) and an estimateq* ofq. Putting dq = aTc − q* gives

m(f,p,aTc) = m(f,p,q* +dq) ≈ m(f,p.q*) +dq m’(f,p,q*) = m(f,p.q*) + (aTc − q*) m’(f,p,q*)

so thataT is now in a linear position with all other terms known, and thus can be analyzed bylinear regression techniques. For more than one parameter the method extends in a direct manner.[5][18][17] After checking that the model has been improved this process can be repeated until convergence. This approach has the advantages that it does not need the parametersq to be able to be determined from an individual data set and the linear regression is on the original error terms[5]

Model validation

[edit]

Where sufficient data is available, division of the data into a separate model construction set and one or twoevaluation sets is recommended. This can be repeated using multiple selections of the construction set and theresulting models averaged or used to evaluate prediction differences.

A statistical test such aschi-squared on the residuals is not particularly useful.[26] The chi squared test requires known standard deviations which are seldom available, and failed tests give no indication of how to improve the model.[11] There are a range of methods to compare both nested and non nested models. These include comparison of model predictions with repeated data.

An attempt to predict the residualsm(, ) with the operating conditionsc using linear regression will show if the residuals can be predicted.[21][22] Residuals that cannot be predicted offer little prospect of improving the model using the current operating conditions.[5] Terms that do predict the residuals are prospective terms to incorporate into the model to improve its performance.[21]

The model inversion technique above can be used as a method of determining whether a model can be improved. In this case selection of nonzero terms is not so important and linear prediction can be done using the significanteigenvectors of theregression matrix. The values inA determined in this manner need to be substituted into the nonlinear model to assess improvements in the model errors. The absence of a significant improvement indicates the available data is not able to improve the current model form using the defined parameters.[5] Extra parameters can be inserted into the model to make this test more comprehensive.

See also

[edit]

References

[edit]
  1. ^abcBohlin, Torsten P. (7 September 2006).Practical Grey-box Process Identification: Theory and Applications. Springer Science & Business Media.ISBN 978-1-84628-403-8.
  2. ^"Grey-box model estimation". Mathworks 2. 2012.
  3. ^Kroll, Andreas (2000). Grey-box models: Concepts and application. In: New Frontiers in Computational Intelligence and its Applications, vol.57 of Frontiers in artificial intelligence and applications, pp. 42-51. IOS Press, Amsterdam.
  4. ^Sohlberg, B., and Jacobsen, E.W., 2008.Grey box modelling - branches and experiences, Proc. 17th World Congress, Int. Federation of Automatic Control, Seoul. pp 11415-11420
  5. ^abcdefghijWhiten, B., 2013.Model completion and validation using inversion of grey box models, ANZIAM J.,54 (CTAC 2012) pp C187–C199.
  6. ^abDraper, Norman R.; Smith, Harry (25 August 2014).Applied Regression Analysis. John Wiley & Sons. pp. 657–.ISBN 978-1-118-62568-2.
  7. ^abWeisberg, Sanford (25 November 2013).Applied Linear Regression. Wiley.ISBN 978-1-118-59485-8.
  8. ^Heaton, J., 2012. Introduction to the math of neural networks, Heaton Research Inc. (Chesterfield, MO),ISBN 978-1475190878
  9. ^Stergiou, C.; Siganos, D. (2013)."Neural networks". Archived fromthe original on 2009-12-16. Retrieved2013-07-03.
  10. ^abLawson, Charles L.; J. Hanson, Richard (1 December 1995).Solving Least Squares Problems. SIAM.ISBN 978-0-89871-356-5.
  11. ^abcPress, W.H.; Teukolsky, S.A.; Vetterling, W.T.; Flannery, B.P. (2007).Numerical Recipes (3rd ed.). Cambridge University Press.ISBN 978-0-521-88068-8.
  12. ^Gelman, Andrew; Carlin, John B.; Stern, Hal S.; Dunson, David B.; Vehtari, Aki; Rubin, Donald B. (1 November 2013).Bayesian Data Analysis, Third Edition. CRC Press.ISBN 978-1-4398-4095-5.
  13. ^Mathworks, 2013.Supported grey box models
  14. ^abHauth, J. (2008),Grey Box Modelling for Nonlinear Systems(PDF) (dissertation,Kaiserslautern University of Technology).
  15. ^abNash, J.C. and Walker-Smith, M. 1987. Nonlinear parameter estimation, Marcel Dekker, Inc. (New York).
  16. ^abcWhiten, W.J., 1971. Model building techniques applied to mineral treatment processes, Symp. on Automatic Control Systems in Mineral Processing Plants, (Australas. Inst. Min. Metall., S. Queensland Branch, Brisbane), 129-148.
  17. ^abcdWhiten, W.J., 1994. Determination of parameter relations within non-linear models, SIGNUM Newsletter, 29(3–4,) 2–5. 10.1145/192527.192535.
  18. ^abcWhiten, B., 2014.Determining the form of ordinary differential equations using model inversion, ANZIAM J. 55 (EMAC2013) pp.C329–C347.
  19. ^Polynomial
  20. ^Spline (mathematics)
  21. ^abcdKojovic, T., and Whiten W. J., 1994. Evaluation of the quality of simulation models, Innovations in mineral processing, (Lauretian University, Sudbury) pp 437–446.ISBN 088667025X
  22. ^abcKojovic, T., 1989. The development and application of Model - an automated model builder for mineral processing, PhD thesis, The University of Queensland.
  23. ^Xiao, J., 1998. Extensions of model building techniques and their applications in mineral processing, PhD thesis, The University of Queensland.
  24. ^abLinhart, H.; Zucchini, W. (1986).Model selection. Wiley.ISBN 978-0-471-83722-0.
  25. ^abMiller, Alan (15 April 2002).Subset Selection in Regression. CRC Press.ISBN 978-1-4200-3593-3.
  26. ^Deming, William Edwards (2000).Out of the Crisis p272. MIT Press.ISBN 978-0-262-54115-2.
Retrieved from "https://en.wikipedia.org/w/index.php?title=Grey_box_model&oldid=1017274915"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp