Escolar Documentos
Profissional Documentos
Cultura Documentos
Contents
Introduction ......................................................................................................2
About This Series ...........................................................................................2
About This Book .............................................................................................2
Supporting Material ........................................................................................3
Optimization – Requirements And Approaches.....................................................4
Basic Definitions.............................................................................................4
“Mathematically Demanding” Approaches ........................................................5
The Need For Alternative Approaches ..............................................................6
Summary ..................................................................................................... 13
Design Improvement – A Designer’s View ......................................................... 15
Concept vs. Existing Design .......................................................................... 15
Design Of Experiment................................................................................... 16
Approximations: Building Better-Behaved Models ........................................... 17
Stochastics................................................................................................... 18
Optimization ................................................................................................ 20
Summary ..................................................................................................... 21
Statistics – A Worms Eye View.......................................................................... 22
Dealing With Populations .............................................................................. 22
Probability ................................................................................................... 25
Experiments................................................................................................. 27
Statistics – A Birds Eye View............................................................................. 34
DOE ............................................................................................................ 34
Variable Screening........................................................................................ 39
Summary ..................................................................................................... 40
Statistics – A Designer’s View ........................................................................... 42
Applications Of Approximate Models .............................................................. 42
Types Of Approximate Models ....................................................................... 47
Optimization ................................................................................................ 49
Reliability ..................................................................................................... 49
Summary ..................................................................................................... 52
HyperStudy - Putting It All Together ................................................................. 54
HyperStudy.................................................................................................. 55
Before The Study ......................................................................................... 55
Performing The Study................................................................................... 57
Reviewing Results ........................................................................................ 61
Glossary And References.................................................................................. 65
1
Introduction CAE and Design Optimization – Advanced
Introduction
About This Series
To make the most of this series you should be an engineering student, in
your third or final year of Mechanical Engineering. You should have access
to licenses of HyperWorks, to the Altair website, and to an instructor who
can guide you through your chosen projects or assignments.
While it’s not essential, a good grasp of the basic principles of statistics and
probability will help you tremendously. Several essential aspects are covered
in this book, although in a qualitative fashion. You may want to treat the
chapter titled Statistics – A Worm’s Eye View as a reference. If you choose
to adopt this approach, at least a cursory reading of this chapter is strongly
recommended.
2
CAE and Design Optimization - Advanced Introduction
The various references cited in the book will probably be most useful after
you have worked through your project and are interpreting the results.
Supporting Material
Your instructor will have the Student Projects that accompany these volumes
– it should certainly be made use of. Further reading and references are
indicated both in this book and in the Instructor’s Manual.
If you find the material interesting, you should also look up the HyperWorks
On-line Help System. The Altair website, www.altair.com, is also likely to be
of interest to you, both for an insight into the evolving technology and to
help you present your project better.
The two volumes of this series that cover manufacturing simulation and
multi-body dynamics complement the techniques covered in this book. They
should be interesting from an application perspective if your areas of work
include non-linear analyses and multi-disciplinary optimization.
George E.P.Box
3
Optimization – Requirements and Approaches CAE and Design Optimization – Advanced
Basic Definitions
A design problem usually involves a set of resources, some limits on these
resources, and one or more performance-criteria to achieve with some
required degree of precision. There is almost always more than one correct
“answer” to a design problem. These are feasible in the sense that they
provide the required functions, and work with available resources. Some
answer are better than others either because they use fewer resources or
because they meet the performance criteria with a higher precision, or with
a higher reliability, or to a higher degree.
The optimum design is one that achieves the function to the best degree
using the fewest resources.
1. The data: things that are given and cannot be changed. For instance, if
a component is to be designed to carry a particular load, the designer
must take the load as “data” and must not seek to change it (unless, of
course, a feasible solution cannot be found). The design space is the
region within which the designer must work. This data is captured in the
analysis model, which could be a Finite Element model or any other
mathematical model.
2. The design variables: things that can be controlled by the designer. For
example, the thickness of sheet-steel or the cross-section of a beam.
4
CAE and Design Optimization – Advanced Optimization – Requirements and Approaches
5
Optimization – Requirements and Approaches CAE and Design Optimization – Advanced
There are some times, however, when linearization is simply not possible.
There are several factors that force us to step beyond linearization and look
for advanced methods of design, analysis and optimization.
6
CAE and Design Optimization – Advanced Optimization – Requirements and Approaches
Multi-Disciplinary Optimization
Almost any problem in the physical world involves multiple effects. Quite
often each of these is characterized by a different equation.
In other cases, we cannot afford to decompose the effects because they are
closely linked. In this case, we try to derive equations that work adequately
to model all effects in a single model. Almost invariably, these equations
are so complex that they require enormous computing resources to
evaluate.
Multi-Objective Optimization
In some cases the quality of the design cannot be
reduced to a single objective. There are multiple measures of product quality
and it’s not obvious which solution is the best.
1
The picture shown is from the Research Quarterly, Summer 2003, of the Los
Alamos National Laboratory.
7
Optimization – Requirements and Approaches CAE and Design Optimization – Advanced
Consider the dilemma a car designer faces. The task may be to design a
vehicle that provides a mileage of at least 20 km / liter and satisfies the
safety statutes.
Yes, that’s certainly possible, but it’s not the best way out. Satisfying the
constraints means you will come up with a design that just satisfies both the
mileage requirement and the safety-statute. It does not mean you will
achieve the best possible design. Your competition may come up with a
design that has an even better mileage or is even safer.
If, as in the case of the vehicle, there are two or more responses to be
minimized, this is called Multi-Objective-Optimization (MOO). Very often the
objectives are opposing. In the car, an increase in safety by adding stiffness
and weight usually results in a fall in fuel economy.
The designer’s task is to deliver the Pareto Frontier to the other members of
the product team, so that an educated decision can be made on what to
trade off.
There may be several designs that exceed both the fuel-economy target and
the statutory safety laws. Deciding which of these is “the” best calls for an
evaluation that lies outside the design engineer’s realm. Most often,
marketing strategies decide this. Choosing a solution invariably involves a
trade-off to arrive at the best compromise.
8
CAE and Design Optimization – Advanced Optimization – Requirements and Approaches
Non-Linear Analyses
Problems that involve large deformations, large strains, plastic flow, contact,
radiative heat transfer, etc. are characterized by non-linear differential
equations. Metal-forming is one such application.
We may not be sure that an optimal solution exists (that is, we may not be
sure that there is a global minimum), but would we be foolish to give up
without a search? In cases like these it makes sense to understand the lay
of the land first.
2
Written by Dave Akins. See http://spacecraft.ssl.umd.edu/ for the complete list.
9
Optimization – Requirements and Approaches CAE and Design Optimization – Advanced
Robust Design
Webster’s dictionary defines a Robust System as one that has
“demonstrated an ability to recover gracefully from the whole range of
exceptional inputs and situations in a given environment. One step below
bulletproof. Carries the additional connotation of elegance in addition to just
careful attention to detail.”
Now recall the definition we used at the beginning of this chapter for
“optimum”: “the greatest degree or best result obtained or obtainable under
specific conditions”.
3
From “Multifactorial analysis of an MOD restored human premolar using auto-mesh
finite element approach”, Lin CL, Chang CH, Ko CC, J Oral Rehabil. 2001
Jun;28(6):576-85
10
CAE and Design Optimization – Advanced Optimization – Requirements and Approaches
The last two words in that definition have often exercised engineers’
ingenuity. Since “all measurements are subject to variation4”, to what extent
can we rely on the existence of “specific” conditions? If the conditions that
are used to arrive at the optimum design themselves are likely to vary, then
how certain can we be that the design is indeed optimal?
How will the design respond to the “exceptional inputs” that Webster
mentions?
Consider the graph, which plots the response vs. a design variable. Let’s
assume the objective is to minimize the response. That is, the best value for
the design variable is the one that gives the least response.
Variations in the design variable (x) result in variations in the response (z).
In the figure, possible variations in x are shown at two points: at the
optimum point and at the “robust” point.
The optimal design is better only if variations in the input can be controlled.
A conscientious designer should specify the maximum permitted variation
4
The observation is credited to W.E.Deming, the importance of whose work on
Quality cannot be overestimated.
11
Optimization – Requirements and Approaches CAE and Design Optimization – Advanced
within which the optimum remains optimal. Else, the claim to be an optimum
design is clearly dubious.
Reliability
Reliability is usually defined as the probability of a measure lying within a
specific range. Almost any product will fail sooner or later. Designs that rely
on large factors of safety are often called conservative. That is, they are
safe, but perhaps excessively so. Can they be made less safe without any
reduction in salability?
Why is this important for our goal of optimizing? We need to recognize that
if an event or condition is probable to occur, not dead-certain, then we can
quantify the probability of failure. In the example of the HP printers, it is
improbable but definitely possible that a customer may pile something on a
printer that is even heavier than the intrepid manager. If the probability of
this is known, and if from this the probability of failure can be estimated,
then the business manager can design a warranty cost accordingly.
5
From an article in Fortune, February 2003.
12
CAE and Design Optimization – Advanced Optimization – Requirements and Approaches
6
Mars probes have been marked by a failure rate that is high by terrestrial
standards, but remarkably low given the lack of information and control. The cost of
failure, unfortunately, is almost always enormous.
13
Optimization – Requirements and Approaches CAE and Design Optimization – Advanced
Robust Design: Not just strong. Flexible! Idiot Proof! Simple! Efficient! A …
high level performance despite … a wide range of changing client and
manufacturing conditions.
Genichi Taguchi
14
CAE and Design Optimization – Advanced Design Improvement – A Designer’s View
15
Design Improvement – A Designer’s View CAE and Design Optimization – Advanced
The rest of this chapter introduces the Advanced Tools listed above: Design
of Experiments, Approximations, Optimization and Stochastic Studies.
Design Of Experiment
The current definition of the scientific approach is often traced back to
Francis Bacon’s procedure “which by slow and faithful toil gathers
information from things and brings it into understanding”. The use of the
adjective “scientific” is taken to mean that there is a verifiable basis for any
assumptions of behavior. Today’s step-by-step approach to understanding
phenomena is
The next sections will cover the use of Design Of Experiments8 (DOE) to
computer simulation, and show why they have been so successful in their
adoption. The use of DOE is so widespread today that we often talk of “the
result of the DOE” rather than the more correct “result of the experiment”!
7
Computers models too can be susceptible to noise. We neglect it in this discussion,
but it is a topic of active research.
8
This method is usually traced back to Ronald Fisher, for whom the F-distribution,
which we will encounter when we discuss ANOVA, is named.
16
CAE and Design Optimization – Advanced Design Improvement – A Designer’s View
Why is this relevant to us? Remember that numerical methods like the Finite
Element Method have been tremendously successful and have, in several
applications, promoted the use of highly-non-linear models. Now recall from
our earlier discussion that the “ill behavior” of some mathematical models
prevents us from using gradient-based optimization methods.
But we cannot use these models in optimization because they are not well-
behaved enough. Either gradients don’t exist, or the model is so demanding
that we cannot afford to evaluate it at enough configurations to search for
the optimum.
17
Design Improvement – A Designer’s View CAE and Design Optimization – Advanced
Stochastics
Most problems in engineering textbooks are very well defined. Since the goal
is usually to acquaint the student with the principles or theories being
taught, complexities are usually ignored. Problems in the real world are
rarely so accommodating. New product designers have to live with constant
changes in specifications, giving rise to the joke that designing to a
specification is like walking on water: easy to do if it’s frozen.
But even if specifications are removed from the list of things that can
change, the engineer in the “real” world is presented with a huge list of
items that cannot be treated as the absolute truth. A spread of 10% is often
considered to be sufficiently accurate for engineering purposes.
Engineers have long recognized and dealt with several forms of variation or
uncertainty. CAD users will be familiar with the stack-up analyses used to
determine tolerances that should be assigned to nominal dimensions. These
are techniques that designers employ to deal with uncertainties in
manufacturing related aspects.
9
Readers of the volume on CAE And Design Optimization – Basics will recognize that
OptiStruct also uses approximate models.
18
CAE and Design Optimization – Advanced Design Improvement – A Designer’s View
Some variations may be entirely random: you may know little of the event,
except that it is possible. Earthquakes are an excellent example of this. No
one, today, can predict when an earthquake can occur. Nor can they predict
what the magnitude of the earthquake will be.
Many items in Mechanical Engineering are stochastic: that is, they are
random, but in a predictable fashion10. You cannot say for certain, but you
can conjecture, or guess, at when the event will occur, and with what
severity.
Usually, stochastic processes are treated as distinct from (and in some sense
the “opposite” of) deterministic processes, which are defined by a set of
equations that specify exactly how the process behaves.
10
All random events can be modeled using stochastic approaches provided there is a
large volume of data.
19
Design Improvement – A Designer’s View CAE and Design Optimization – Advanced
It’s important to remember that this does not mean you should neglect
factors of safety. Failure prevention is an essential part of design. Stochastic
behavior comes in particularly useful when we want to quantify the effects
of variations on the design. We can either vary the controllable parameters,
or investigate the effects of variations of uncontrollable parameters.
In other words, either we are already aware that it won’t fail and we want to
see how much we can tweak it towards better performance, or we are
willing to tolerate some failure in return for better performance.
Optimization
DOE, Approximate Models and Stochastic Models are all ways to investigate
the behavior of our Design Space. We can use these investigations to
improve our design in several ways.
We can also perform trade-off studies to identify Pareto Optimal designs and
construct Pareto Frontiers.
In cases of MDO, we may also need to work with multiple solvers. The
variables of interest may have different behaviors in each “physics” and we
20
CAE and Design Optimization – Advanced Design Improvement – A Designer’s View
need a method to search for the solution that is the best across all these
multiple-disciplines.
Summary
To optimize design problems that demand the use of complex mathematical
models, or multiple disciplines, or that have multiple objectives, then, our
requirements of computer software are:
• The ability to build models for and extract results from multiple
solvers
After all, facts are facts, and although we may quote one to another with a
chuckle the words of the Wise Statesman, “Lies - damn lies - and statistics,”
still there are some easy figures the simplest must understand, and the
astutest cannot wriggle out of.
Lord Courtney
21
Statistics – A Worm’s Eye View CAE and Design Optimization – Advanced
This chapter lists terms that we will encounter frequently in our usage of
CAE for advanced design optimization. Most are defined without any
preamble or context. These will be provided in the next two chapters.
11
Quoted by E.T.Bell, in “Men of Mathematics”
22
CAE and Design Optimization – Advanced Statistics – A Worm’s Eye View
A good example of this would be the population of a small village vs. the
population of a large country. Since the former is likely to be a few hundreds
at the most, data can be easily manipulated by treating each inhabitant as a
distinct member of the village. In the latter case, where the population lies
in the hundreds of millions, it is much easier to treat it as a continuous set.
In this case, all the theorems of differential and integral calculus can be
applied.
Random Variable
We use an algebraic symbol to represent any member of the population.
This symbol is called a variable. It can take on the characteristics of any
member of the population, just as a variable in an algebraic equation can
take on any value in the range / domain.
There is little point in working with variables that are not random, just as
there is little point in playing a game where the die is loaded: the outcome is
a foregone conclusion. Therefore we focus our attention on random
variables only.
Discerning a trend, of course, is the main reason we collect the data in the
first place: as with most applications of modern science, the goal is to
understand the behavior of the population. Based on this understanding,
predictions can be made and comparisons drawn.
Even though the variable is random, statistical studies show that values tend
to cluster around the average. Measures of this tendency to cluster are
important in our efforts to characterize populations.
23
Statistics – A Worm’s Eye View CAE and Design Optimization – Advanced
Measures Of Central Tendency
There are three basic measures: the mean, the median, and the mode.
While there are several measures of the mean, most often we look for a
value that can be interpreted as the average value.
While the mean, median and mode are often distinct from each other, they
can sometimes be the same. In particular, for the Normal Distribution that is
described below and that is our main focus of attention, all three are the
same. The most common symbol for the mean is µ, while x is sometimes
used.
Standard Deviation
While measures of central tendency indicate where the mean lies, how can
we measure the spread of the data? This is done using the standard
deviation, for which the usual symbol is σ. It is a standard measure of the
dispersion of the data from the mean.
Variance
The variance is nothing but the square of σ. It represents the average of the
squares of deviations from the mean. Squaring the deviations prevents
negative values from canceling out positive values.
Coefficient of Variance
The variance or standard deviation by themselves may provide a misleading
picture. For instance, consider a population with a standard deviation of 3.75
and a mean of 1,000. If we want to compare it with another population
whose standard deviation is also 3.75, but whose mean is 10, it is “obvious”
that in the former case the data is more tightly clustered around the mean.
In mathematical terms, “obvious” means that we are normalizing the
standard deviation with respect to the mean.
24
CAE and Design Optimization – Advanced Statistics – A Worm’s Eye View
Probability
To define probability, it’s hard to improve on Aristotle, who said “the
probable is what usually happens.”
Continuous Distributions
When we work with populations, we find it useful to characterize the
variation of the random variable over the entire population in terms of a
distribution. That is, a symbolic form or formula that we can use to calculate
probabilities of occurrence of various values. There are several different
distributions that have been encountered in scientific studies: Binomial,
Normal, Weibull, Poisson, etc.
−( x− µ )2
f ( x) = 1
e 2σ 2
2πσ 2
In this equation µ is the mean and σ is the standard deviation. The plot of
this Probability Density Function (PDF) is the reason it is also called the Bell-
Shaped Distribution.
25
Statistics – A Worm’s Eye View CAE and Design Optimization – Advanced
CDF
The integral of the PDF is called the Cumulative Density Function, since it
represents the sum of probabilities of all values that lie within the limits of
the integral.
x
F ( x) = ∫ f (t )dt
−∞
This is useful.
+∞
∫ f (t )dt = 1
−∞
12
Obviously, it is easier to look up a table than to evaluate the integral! In fact, the
integral cannot be evaluated analytically.
26
CAE and Design Optimization – Advanced Statistics – A Worm’s Eye View
Experiments
Even with discrete variables, enumerating all possible values is often beyond
the power of available resources. For continuous distributions, it is obviously
out of the question. Data is sometimes collected by observation. The
observer makes no effort to control the variables, restricting effort to
collecting data and drawing conclusions from these. If an independent
variable does not change during the period of observation, the analyst can
draw no conclusions about its effect on behavior of the population.
13
Finite Element beginners will sympathize: the choice of elements is often just as
confusing for a beginner.
27
Statistics – A Worm’s Eye View CAE and Design Optimization – Advanced
Random Samples
In order to do this effectively, the sample must be free of bias. That is, it
must not be predisposed to any particular sets of values.
Samples that are free of bias are called random samples. Deciding which
parts of the population to sample is an important question. Indeed, “the
generation of random numbers is too important to be left to chance14.”
As mentioned above, the Central Limit Theorem tells us that the properties
of samples follow a Normal Distribution even if the population itself does
not. Why is this important to us?
20
∑µ i
s= i =1
20
Because the Central Limit Theorem tells us that the means of the samples
themselves follow a Normal Distribution, regardless of the distribution the
parent population follows, we can use the properties of the Normal
Distribution to calculate a confidence interval. This tells us the probability
that the mean of the population will lie within some specific distance from
the sample-mean. For example, there is a 95% probability that the mean of
the population will lie within the 95% confidence interval of the sample
mean.
14
Robert Coveyou, of the Oak Ridge National Laboratory, quoted by Ivars Peterson
in “The Jungles of Randomness; A Mathematical Safari”
28
CAE and Design Optimization – Advanced Statistics – A Worm’s Eye View
The smaller p is, the higher the confidence in the estimated values. p = 0.05
is the same as a 95% confidence, while p = 0.01 is the 99% confidence
interval. If a report says that the “95% confidence interval” for a response R
is 10.5 to 21.27, this means that there is a 95% probability that R will lie
between 10.5 and 21.27.
ANOVA
To make a meaningful comparison between measurements of different
samples, the conditions should be the same for all samples. This is obvious.
The control variables are called factors. Values that the control variables can
take are called levels. The values of interest are called responses. Let’s say
the experimenter has 1 factor, which can take 3 levels. And let’s say the
experimenter has constructed 5 samples (that is, 5 groups).
29
Statistics – A Worm’s Eye View CAE and Design Optimization – Advanced
Then for each response we can present the results of the experiment as in
the table below:
Note that R11 would have been the same as R21 in the absence of noise.
Also, R12 would have been the same as R11 if the level had been the same.
Without going into the equations, we will state simply that ANOVA involves
calculating variances in the response between columns (between groups)
and between rows (within groups). Then, using the F-distribution, the
significance of the difference is estimated15.
The results of the Anova are often displayed in a table showing the factors
and the levels of significance.
ss df MS F p
Factor 1
Factor 2
Factor 3
Factor 4
Error
Total SS
15
For details on the F-Distribution and why it is used here, look up the references.
30
CAE and Design Optimization – Advanced Statistics – A Worm’s Eye View
If there were multiple responses, ANOVA would involve drawing up the table
above for each response.
If there are multiple variables in the experiment, then the term MANOVA is
sometimes used.
This is called the Monte Carlo approach. The main advantage of the Monte
Carlo methods is that they are scaleable; they can be applied as easily to a
1000-dimensioned problem as to a 1-dimensioned problem. In engineering
practice, these methods are often deployed when there are a large number
of variables.
There are several variations on the method, all of which are sometimes
referred to collectively as the Monte Carlo methods, since the principle
behind all of them is the same.
When dealing with multiple variables, each of which can have multiple
levels, the problem of designing an experiment itself can become intractable.
As we will see in the next Chapter, designing a meaningful experiment is not
a trivial task!
31
Statistics – A Worm’s Eye View CAE and Design Optimization – Advanced
Uni-variate Analyses
If a distribution has only one independent variable, we say it is univariate.
From the available data (gathered either through experiment or by
observation) we calculate the correlation coefficient between the variable
and the response.
32
CAE and Design Optimization – Advanced Statistics – A Worm’s Eye View
Multi-variate Analyses
If there are multiple independent variables, as is normally the case, we use
multi-variate statistics. In this case, the regression equation is a “surface”
instead of a curve.
However, factors are often linked. Some may be independent, some may
work in tandem, some in opposition. One of the jobs of the experimenter is
to search for such linkages. The OFAT approach will not uncover linkages
between variables. An alternate approach, called DOE (described in the next
chapter) is better suited to the task.
Augustus De Morgan
33
Statistics – A Bird’s Eye View CAE and Design Optimization – Advanced
We saw how levels of factors are varied across samples, responses are
measured, and the significance of the measurements is estimated. All these
are based on “sound” mathematical principles. That is, the assumptions and
proofs are well laid out.
There is one important question that the previous chapter did not address.
How are the samples themselves to be constructed?
DOE
Design Of Experiments is the procedure of selecting the points in the design
space where responses are to be measured.
• Factor 2 can have any of 4 levels – L1, L2, L3, and L4. These
represent the values 0, 2, 5 and 10.
Temperature
420° 450° 470°
0
passes
# of
2
5
10
As the table shows, there are 12 combinations of factors that are possible –
that is, 12 feasible process-settings (3*4, since there are 3 possible levels
for the temperature and 4 for the number-of-passes).
35
Statistics – A Bird’s Eye View CAE and Design Optimization – Advanced
First, it brings in a new example for the “population”. In this case, the
population consists of all the possible combinations of values of all the
factors. Since the factors are discrete, a census is possible: that is, evaluate
at all possible combinations and choose the best.
Note that if we had had access to 100 identical mills, we could have
performed 10 tests, each on a different mill. Clearly this would define a
different “population” since the “mill being tested” would be a variable in
this case.
Full Factorial
An approach where all possible combinations of levels and factors is
evaluated (that is, the responses are measured) is called a full factorial
experiment. Obviously, it can increase in size very rapidly. For 20 factors and
3 levels per factor, the number of combinations is 320, which is roughly
3x109.
In general, if there are n factors, and each factor can have li levels, then the
total number of combinations in a full-factorial experiment is given by
∏l
i =1
i
36
CAE and Design Optimization – Advanced Statistics – A Bird’s Eye View
Such experiments are called 2-level experiments, and are usually used to
screen variables – to eliminate the trivial and retain the important ones.
Fractional Factorial
In this design, we choose a fraction of the full-factorial. This fraction can be
½, ¼, etc. For instance, a 2-level 3-factor design has 8 combinations if a
full-factorial design is chosen. The full factorial design is called a 23 design.
If we choose to run a ½ factorial, we call it a 23-1 design, and it will contain
4 measurements.
A full discussion of how to choose the fraction is beyond the scope of this
book. Look up one of the References for this detail, as well as for a
discussion on which subset of the full-factorial to choose for a given fraction.
Standard literature often describes fractional-factorial designs as 2(k-p), since
there are usually 2 levels per factor. For example, a ½ factorial design with
6 factors is called a 2(6-1) design. “k” is “6”, of course, because there are 6
factors. “p” is “1” because this is a “½ factorial design”. If it were a ¼
factorial design, “p” would be 2.
The tables below show the 23 full factorial, and one possible ½ fractional
factorial. In these tables, we have chosen +1 and –1 to represent the
possible levels for each factor.
37
Statistics – A Bird’s Eye View CAE and Design Optimization – Advanced
Full Factorial ½ Fractional Factorial
F1 F2 F3 F1 F2 F3
1 +1 +1 +1 1 -1 -1 +1
2 +1 +1 -1 2 +1 -1 -1
3 +1 -1 +1 3 -1 +1 -1
4 +1 -1 -1 4 +1 +1 +1
5 -1 +1 +1
6 -1 +1 -1
7 -1 -1 +1
8 -1 -1 -1
The fractional factorial design was generated by first choosing 4 values for
F1 and F2. The levels for F3 were chosen by multiplying the chosen levels of
F1 and F2. We say that we have “confounded F3 by F1 and F2”.
Fractional Factorial designs are suitable when some factors are considered
more important than others. We are willing to give up some resolution in the
weaker factors in return for the economy we gain by virtue of the shorter
experiment. However, remember that there is a-priori judgment involved in
deciding which factors to treat as “weaker”.
Since the effects of the lesser factors is reduced, such designs are
sometimes called screening designs.
Plackett-Burman
Named for the R.L.Plackett and J.P.Burman16, this design reduces the
number of runs even further. The main effects are heavily confounded. For
16
They published their paper titled “The Design Of Optimal Multifactorial Designs” in
1946
38
CAE and Design Optimization – Advanced Statistics – A Bird’s Eye View
Central Composite
Also called Box-Wilson designs, there are several variations of this method:
circumscribed, inscribed, and face centered. The first two require 5 levels
per factor, while the third requires 3. CC designs can be full-factorial or
fractional-factorial.
Remember that a 2-level experiment can only capture linear effects, while a
3-level experiment can capture quadratic effects, and so on.
Box-Behnken
A Box-Behnken design, also called a quadratic design, is a slightly more
economical variation of the full-factorial CC design but can be more
expensive than a fractional-factorial CC Design.
Other Designs
There are several other types of designs, such as the Latin Hypercube
Design, Taguchi Design, and so on. Some designs are particularly suited for
computer experiments.
Variable Screening
In many cases, we want to check whether selected factors have an effect on
the responses or not. To do this, multiple samples are constructed, and the
experiment is repeated across these samples. As described in the previous
chapter, ANOVA is used to test levels of significance of the various factors. If
the effect of the variation of a factor is significantly more important than the
change of groups, then it is retained. Else it is screened.
39
Statistics – A Bird’s Eye View CAE and Design Optimization – Advanced
Computer Experiments
In many cases, the experiment is conducted using a computer model. This,
of course, is of particular interest to us.
Summary
The previous chapter covered essential terms in statistics and probability.
This chapter put those terms together to explain how experiments are
conducted, and outlined the issues involved in the design of experiments.
17
As pointed out earlier, computers models too can be susceptible to noise.
Neglected in our discussion, as it is a different subject.
40
CAE and Design Optimization – Advanced Statistics – A Bird’s Eye View
The next chapter looks at the issue from an engineering design point of
view, bringing together terms and techniques relevant to CAE.
However beautiful the strategy, you should occasionally look at the results.
Winston Churchill
41
Statistics – A Designer’s View CAE and Design Optimization – Advanced
First, we’ll list the advantages Approximate models have. Different scenarios
may benefit from one or more of these. It is possible that a scenario may
not need any of these benefits, in which case Approximate Models can be
dispensed with completely.
42
CAE and Design Optimization – Advanced Statistics – A Designer’s View
For cases like these, approximate models offer a good way out. We choose
the form of the approximate model to ensure that it is differentiable or
otherwise well-behaved. (Note that differentiability is important for gradient-
based optimization methods, but is not required for other design
improvement methods.)
What’s more, they offer a way out of the second problem too. If you find an
error in the experiment-design, you can repair the approximate model:
points that define the model can be added, removed or moved.
18
“Failure Analysis of Rapid Prototyped Tooling in Sheet Metal Forming – Cylindrical
Cup Drawing”, Y.Park and J.S.Colton, Transactions of the ASME, Vol 127, February
2005
43
Statistics – A Designer’s View CAE and Design Optimization – Advanced
Variable Screening
Testing takes time and effort. It is expensive. The more the factors you
want to test, the higher the time and expense. What if some factors are
unimportant? Can you conduct a preliminary investigation to rank the
importance of various factors? Can you then save time and money by
excluding the lower-ranked ones from more detailed studies?
Everybody believes
2. Mathematical Models, which usually involve some
test results except
approximation again, are used to reflect the observed the test engineer.”
behavior. These are not always well-behaved, but are M.Racicot
often called high fidelity models.
In view of the 3 steps listed above, let us state the question more precisely,
recognizing that there are actually two different questions:
44
CAE and Design Optimization – Advanced Statistics – A Designer’s View
In the second case, we have a computer model that is tried and tested.
There is no doubt about its validity. This is the high fidelity model. It could
be an analytical expression or a numerical model19. However the high fidelity
may have several input parameters. If we are to use it to conduct
experiments, which of these variables should we include in the experiment?
If numerical models such as FEM are used, there is a model that reliably
calculates response from inputs, but is not analytical. So sensitivity must be
calculated numerically.
The first case (screening between observed data and the high-fidelity
model) is addressed by parameter identification. In this approach, the
results of a physical experiment are set up as target values. The computer
model is run with various levels of many factors. By inspecting the
19
Transfer Functions (covered in most courses on Control Systems) provide excellent
examples of high-fidelity analytical models. Many linear processes can be accurately
described by numerical models.
45
Statistics – A Designer’s View CAE and Design Optimization – Advanced
Screening experiments typically involve only two levels for each factor. The
designer is encouraged to include as many factors as possible. ANOVA is
conducted on the factors themselves to quantify their effects on the
responses.
20
For an excellent description see “Automotive crashworthiness design using
response surface-based variable screening and optimization”, K.J.Craig, N.Stander,
D.A.Doorge, S.Varadappa, International Journal for Computer-Aided Engineering
Software, Vol 22 No.1, 2005, pp.38-61
46
CAE and Design Optimization – Advanced Statistics – A Designer’s View
Unlike the earlier example of ANOVA, the results of this screening are
usually presented in graphical form. If the approximate model is to be used
calculate multiple responses, one graph is presented for each response. For
instance consider the histograms below21, in which the length of each bar
indicates the effect of the corresponding factor.
Remember that the estimated effect has some error. This error is calculated
by the ANOVA. The F-values are used to estimate the “confidence” in the
estimate. This is usually taken to be 95%. In the graphs below, the lighter
part of each bar is that part of the effect that the analysis is 95% confident
of. The darker part, which is the lower-confidence fraction of the total effect,
is treated as error.
21
From “Automotive crashworthiness design using response surface-based variable
screening and optimization”, cited above.
47
Statistics – A Designer’s View CAE and Design Optimization – Advanced
The least-squares approach means that the surface is unlikely to match the
high-fidelity solution anywhere. A large number of sampling points and a
higher order of polynomial helps improve accuracy of the response surface.
This method is usually good at capturing global minima, since it tends to
smooth out local minima.
Kriging
Named for D.G.Krige, who was trying to determine the grade of ore from
samples, this method is sometimes preferred because of its improved
accuracy. Unlike the least-squares fits, the surface interpolates the values at
the sampling points.
48
CAE and Design Optimization – Advanced Statistics – A Designer’s View
Optimization
Remember that the approximate model is not essential. If the high-fidelity
model is well-behaved and computationally efficient, optimization can be
performed without using an approximate model.
Reliability
Recall our earlier definition of reliability: the probability of a measure lying
within a specific value. One of the main drawbacks of an “academic”
knowledge of engineering is that most textbooks present well-defined
49
Statistics – A Designer’s View CAE and Design Optimization – Advanced
22
In his report on the Shuttle disaster of 1986. It makes excellent reading for
engineers! See http://science.ksc.nasa.gov/shuttle/missions/51-l/docs/rogers-
commission/Appendix-F.txt for details.
50
CAE and Design Optimization – Advanced Statistics – A Designer’s View
However, remember that cost is not the only reason to adopt DFSS. The
impossibility of eliminating variations in data means many problems cannot
be treated as purely deterministic.
A Design Engineer using CAE has to accept the fact that there will be
variations in any of
• Material properties
• Loads
• Boundary conditions
• Initial conditions
• Geometry errors
• Assembly errors
• Solver precision
• Choice of model (mesh, element type, algorithm, etc.)
Noting that the aerospace industry emphasizes safety over cost, an extract
from an aerospace-application conference23 presentation serves to highlight
the fact that cost is not the only motivation for stochastic analyses:
23
Klein M., Schueller G., “Proabilistic Approach to Structural Factors of Safety in
Aerospace”, Proceedings of the CNES Spacecraft Structures and Mechanical Testing
Conference, Paris, June 1994
51
Statistics – A Designer’s View CAE and Design Optimization – Advanced
Thermal Protection In-plane Tension 12-24%
In-plane Compression 15020%
Load Scatter
Load Type Origin Of Results CV
Launch vehicle thrust STS, Ariane 5%
Launch vehicle quasi- STS, Ariane, Delta 30%
static loads
Transient Ariane 4 60%
Thermal Thermal Tests 8-20%
Deployment Shocks (solar Aerospatiale 10%
array)
Thruster Burn Calibration Tests 2%
Acoustic Ariane 4 and STS (flight) 30%
Vibration Satellite Tests 20%
We have already seen the issues related to the use of high-fidelity models,
and have built methods to estimate the error if we use approximate models
both to explore the design space (using DOE) and to optimize.
These analyses are usually done using Monte Carlo methods. An enormous
number of runs can be required to make the best use of statistical effects.
Monte Carlo experiments frequently require thousands of runs.
Summary
The first chapter discussed MOO, MDO, and non-linear models, and
introduced the need for experiments.
52
CAE and Design Optimization – Advanced Statistics – A Designer’s View
• estimate reliability
• perform trade-offs
You'll never have all the information you need to make a decision. If you did,
it would be a foregone conclusion, not a decision.
David Mahoney
53
HyperStudy - Putting It All Together CAE and Design Optimization – Advanced
• the ability to build models for and extract results from multiple
solvers
54
CAE and Design Optimization – Advanced HyperStudy - Putting It All Together
HyperStudy
With direct connections to Altair’s Solvers24, interfaces to several other
solvers, and the ability to interface with any solver, HyperStudy is a good
way to achieve most of the requirements we’ve studied so far.
If you have set up a model in HyperMesh, for example, you can invoke
HyperStudy which then has direct access to model variables.
If you’re using a solver that is not from Altair but that interfaces with
HyperMesh and HyperView, then the approach is a little different. You would
first create a data file for the solver (you could use HyperMesh to do this,
though it’s not essential). Then you would read this data file in HyperStudy,
and set up your study by choosing variables from the data file.
HyperStudy uses its interfacing abilities both to generate additional data files
for the solver (with changed data) for each experimental measurement, and
to read in evaluated responses.
The assignments illustrate the steps in detail, so they will not be covered in
this chapter. Here, we will look at the steps you should take to setup and
review a problem. With the background laid out in the earlier chapters, you
should be able to follow why we take the approaches described below. You
should pay particular attention to the recommended methods to review
results.
24
Of which Finite Element analyses, kinematic and dynamic analyses of mechanisms,
and sheet-metal forming simulation are covered by the other volumes of this series.
25
This is covered in the volume titled Managing The CAE Process
55
HyperStudy - Putting It All Together CAE and Design Optimization – Advanced
Even before you start building your model, you should be clear on the
answers to the following:
Also spend some time planning how you will collate and present results. FE
models are often interpreted by displaying results on the 3D model of the
component, but this may not be appropriate. Remember that you will have
to review a much larger magnitude of results: you may have results of
several dozen FE analyses!
Also, there are numerous other forms of simulation and applications that do
not use the kind of graphics that FEA uses. A polynomial transfer function
may be the high-fidelity model!
56
CAE and Design Optimization – Advanced HyperStudy - Putting It All Together
Study Setup
A “study” is saved in an XML26 file. Unless you’re working with a data
management tool (such as Altair Data Manager) it’s a good idea to create a
separate folder for each study.
3. The responses. These are chosen from the results of the evaluation
carried out by the high-fidelity model. For instance, the stress at a node
can be selected either from a text-output file or a binary-results file.
26
Short for Extensible Markup Language. Uses tags, similar to HTML files.
57
HyperStudy - Putting It All Together CAE and Design Optimization – Advanced
DOE Study
Once the study has been defined, you specify what type of experiment you
want to use.
Remember that
For the study, you will need to distinguish between controlled and
uncontrolled variables. The former consist of design variables that you want
to manipulate as a part of your design, while the latter are due to
uncontrollable noise. You can use different DOEs for the controlled and
uncontrolled variables, depending on the amount of effort you can afford to
spend.
58
CAE and Design Optimization – Advanced HyperStudy - Putting It All Together
Once you have done this, you fine-tune your experiment. Recall the table
drawn up earlier when the Fractional Factorial design was discussed. Most
DOEs are specified as a matrix, showing the levels that will be used for each
factor, and which effects will be confounded. Fine-tuning means you can
edit the matrix to change the allocations for each factor / level and the
interactions between them.
Once the DOE has been specified, HyperStudy runs the analyses and
extracts responses.
Approximation
If the DOE is to be followed by an optimization or a stochastic analysis, it’s a
good idea to build an approximation. Even if these are not planned, the use
of approximations for variable screening is a very useful insight.
You can define an approximation for any or all responses calculated by the
DOE. Each response can have a different type of approximation. You can
also build multiple approximations for a each response, which you would do
if you are unclear on which is best suited.
The DOE points are now used to construct the approximation. You can use
some of the points to define the approximation, and others to “validate” it.
If, for instance, you have conducted 6 runs, you may use runs 1/3/5/6 to
create the approximation. Then you can check the efficacy of the
approximation by using runs 2/4 to validate it and calculate the residuals.
Optimization Study
Remember that an optimization is not the only reason HyperStudy is used.
Therefore it’s logical that none of the previous steps defined any of the
optimization-specific terms: constraints and objectives.
Constraints are limits on the design variables – any design that crosses
these limits is infeasible. Objectives are the quantitative measures of design
59
HyperStudy - Putting It All Together CAE and Design Optimization – Advanced
In any case, you can choose to use either the high-fidelity model or the
approximate model (provided one has been built, of course).
You will want to be judicious in your choice of iterations for the analysis. It is
better to use a small limit to start with. Even if this is unsuccessful, you can
use it to see how the optimization is progressing. Subsequent optimizations
can restart from this analysis, meaning the initial investigation is not wasted.
Stochastic Study
A stochastic study, which generated the PDF of responses based on PDFs of
design variables, can be performed directly after the Study Setup is
complete. Normal, Uniform, Triangular, Exponential and Weibull distributions
are supported.
A DOE is not required, since the sampling method is specified here: you can
choose between
60
CAE and Design Optimization – Advanced HyperStudy - Putting It All Together
Once this is done, HyperStudy evaluates the responses using either the
high-fidelity model or the approximation, depending on your choice. That is,
you explicitly tell HyperStudy which to use.
Reviewing Results
Statistics are notorious for their ability to allow the analyst to draw a variety
of different conclusions from the same data. Before discussing what to
review and how, it’s useful to summarize the motivations for each type of
study.
DOE
The goals usually are one or more of the below:
Approximation
The main reasons we choose to build approximate models are:
61
HyperStudy - Putting It All Together CAE and Design Optimization – Advanced
Optimization
An optimization is just a way to locate a particular point on the response-
curve or surface. The two reasons we search for such a point are
Stochastic Study
While every problem is non-deterministic in reality, remember that it is often
possible to get a good answer using a deterministic model and applying an
appropriate factor of safety. It’s when the factor-of-safety approach is either
too expensive or covers up too much detail that we turn to stochastic
analyses. The main goals are:
The plot on the right shows a negative correlation. As the design variable
increases, the response decreases. If the points were distributed more or
less parallel to the x-axis, we would conclude that the response is
independent of the design variable.
62
CAE and Design Optimization – Advanced HyperStudy - Putting It All Together
The plot on the left shows a variation, but no discernible pattern. In all
likelihood, there is another factor that is causing the change in the response.
Histograms are bar-charts, with data clubbed into “classes”. Classes are also
called “groups”, “buckets” or bins. In general, the more the bins, the finer
the resolution of the data. Of course, you need to have a large number of
runs for the bins to be high.
HyperGraph
HyperStudy provides quick and easy graphical display of most data, but
there are times when you may want to generate different plots or forms of
display of data. HyperGraph provides support for direct import of data from
other HyperWorks applications, including HyperStudy.
63
HyperStudy - Putting It All Together CAE and Design Optimization – Advanced
You got to be very careful if you don't know where you're going, because you
might not get there.
Yogi Berra
64
CAE and Design Optimization – Advanced Glossary and References
Ant-hill Plot Also called a scatter plot, this is a graph showing the change in one
variable as the other changes. It is used to look for a correlation
between the variables.
Confidence The interval within which a variable may lie with a particular
confidence. A 95% confidence is often used in engineering
Interval
applications.
Convex Function A function that has only one minimum in the domain. This minimum is
the “global minimum”.
Dependent The responses that are measured by the experimenter, who varies the
dependent variables as per a plan. The plan is called the DOE.
Variable
Global Variable If an analysis involves multiple solvers (as in MDO), a global variable is
one that is relevant in all contexts.
Interaction Means that two or more independent variables are linked, not
independent of each other. This does not mean there’s a cause-and-
effect relationship. It only means that changing one means the other
65
Glossary and References CAE and Design Optimization – Advanced
effect relationship. It only means that changing one means the other
changes too. If an interaction is detected, further study is required to
understand if there’s a cause-and-effect linkage between these two, or
perhaps a third independent variable that’s the cause.
Latin Square A square array in which each letter or symbol appears exactly once in
each row and column. Used to design experiments by treating the
different rows as levels of the first factor, the columns as levels of the
second factor and the contents of each cell of the array as levels of
the third factor.
Level Values that a factor can take. Can be qualitative (“good” or “bad”) or
quantitative. In CAE, levels are quantitative.
Orthogonal Array In the context of DOE, refers to a table of rows and columns such that
for any pair of columns (i.e. factors) all combinations of levels occur,
and all occur the same number of times.
Outlier A point in the design space that does not follow the general pattern.
This may be either due to noise, in which case the point should be
ignored, or due to actual behavior of the system. In the latter case,
the designer will need to decide how this affects design decisions.
Pareto Frontier Relevant to MOO. Plot showing Pareto Optimal designs. Helps choose
the best compromise, often on non-engineering bases.
Regression Usually linear, it’s an equation that links two correlated variables. We
calculate the regression line only after calculating the correlation
coefficient and ensuring that the variables are indeed correlated.
66
CAE and Design Optimization – Advanced Glossary and References
coefficient and ensuring that the variables are indeed correlated.
Residual The difference between observed and predicted values. The smaller
the residual, the better the model used for prediction. For instance,
you will want to check the difference between the approximation and
the high-fidelity model, to check whether you can reliably use the
approximation.
Measures the goodness-of-fit.
Resolution Measure of the ability of a DOE to capture interactions. The higher the
resolution, the better the capability. A full-factorial design has an ∞
resolution. Confounding reduces the resolution of an experiment. In
practice, a resolution of 5 is excellent, while a resolution of 3 is
sufficient for screening.
Saturated Design A DOE in which the number of evaluation points equals the number of
unknown coefficients in the approximation. It is not possible to test for
a lack of fit.
Stochastic Something that involves chance or probability, but with an overall and
measurable trend or direction – this make sit possible to predict the
behavior. Engineers frequently encounter stochastic processes and
stochastic variables.
Worst Case A formulation in which the objective is minimized with respect to some
variables and maximized with respect to others. This is not the same
Design
as a Min-Max design, where the minimization or maximization is done
with respect to the same variables.
References
Hill, T. & Lewicki, P. (2006). Statistics Methods and Applications.
(http://www.statsoft.com/textbook/stathome.html)
Other Resources
www.altair-india.com/edu, which is periodically updated, contains case
studies of actual usage. It also carries tips on software usage.
68
CAE and Design Optimization – Advanced Glossary and References
0.6, while the cell marked with a “$” (0.7357) is the probability that z will be
< 0.63.
Using Excel, the formula NORMSDIST(0.63) gives you 0.7357. For a general
normal distribution with a mean µ and a standard deviation σ, the formula
NORMDIST(0.63, µ, σ, TRUE) gives 0.7357. To do this manually, convert the
random variable to the standard form by subtracting the mean and dividing
by the standard deviation. That is,
x−µ
z=
σ
Then use this transformed variable z with the above table to find P(z).
You can, of course, use the Normal Table backwards: that is, you can look
up the value that z should have for, let’s say, a 95% probability of
occurrence. Remember that the values above are cumulative densities. That
is, they represent the integral from –∞ to the given value.
69
Glossary and References CAE and Design Optimization – Advanced
aX = N(aµ1, a2σ12)
The F Distribution
This is the distribution of the ratio of two estimates of the variance. It
results when variables that follow a normal distribution are sampled, and the
measured values are squared and summed. The distribution depends on the
“degrees of freedom”: that is, the number of levels (within groups) and the
number of samples (between groups). The variability within a group is
assumed to occur because of error, while the variability across groups is
assumed to occur because of true variance and error. ANOVA is a way to
separate the error from the true variance.
70