Brigham Young University BYU ScholarsArchive All Theses and Dissertations 2011-07-29 Efficient, Accurate, and Non-Gaussian Error Propagation Through Nonlinear, Closed-Form, Analytical System Models Travis V. Anderson Brigham Young University - Provo Follow this and additional works at: http://scholarsarchive.byu.edu/etd Part of the Mechanical Engineering Commons BYU ScholarsArchive Citation Anderson, Travis V., "Efficient, Accurate, and Non-Gaussian Error Propagation Through Nonlinear, Closed-Form, Analytical System Models" (2011). All Theses and Dissertations. Paper 2675. This Thesis is brought to you for free and open access by BYU ScholarsArchive. It has been accepted for inclusion in All Theses and Dissertations by an authorized administrator of BYU ScholarsArchive. For more information, please contact [email protected]. Efficient, Accurate, and Non-Gaussian Statistical Error Propagation Through Nonlinear, Closed-Form, Analytical System Models Travis V. Anderson A thesis submitted to the faculty of Brigham Young University in partial fulfillment of the requirements for the degree of Master of Science Christopher A. Mattson, Chair David T. Fullwood Kenneth W. Chase Department of Mechanical Engineering Brigham Young University Thesis Completed July 2011 c 2011 Travis V. Anderson Copyright All Rights Reserved ABSTRACT Efficient, Accurate, and Non-Gaussian Statistical Error Propagation Through Nonlinear, Closed-Form, Analytical System Models Travis V. Anderson Department of Mechanical Engineering, BYU Master of Science Uncertainty analysis is an important part of system design. The formula for error propagation through a system model that is most-often cited in literature is based on a first-order Taylor series. This formula makes several important assumptions and has several important limitations that are often ignored. This thesis explores these assumptions and addresses two of the major limitations. First, the results obtained from propagating error through nonlinear systems can be wrong by one or more orders of magnitude, due to the linearization inherent in a first-order Taylor series. This thesis presents a method for overcoming that inaccuracy that is capable of achieving fourth-order accuracy without significant additional computational cost. Second, system designers using a Taylor series to propagate error typically only propagate a mean and variance and ignore all higher-order statistics. Consequently, a Gaussian output distribution must be assumed, which often does not reflect reality. This thesis presents a proof that nonlinear systems do not produce Gaussian output distributions, even when inputs are Gaussian. A second-order Taylor series is then used to propagate both skewness and kurtosis through a system model. This allows the system designer to obtain a fully-described non-Gaussian output distribution. The benefits of having a fully-described output distribution are demonstrated using the examples of both a flat rolling metalworking process and the propeller component of a solar-powered unmanned aerial vehicle. Keywords: uncertainty analysis, statistical error propagation, system modeling, Taylor series expansion, variance propagation, skewness propagation, kurtosis propagation ACKNOWLEDGMENTS I would like to thank Dr. Christopher Mattson for supporting me in this research. He has helped open my eyes to the world of research. He has pushed me to explore the field of engineering design and find solutions to some problems in statistical uncertainty analysis. He has been a great mentor to me. I would like to thank Dr. David Fullwood for providing his insights and lending his mathematical expertise to furthering this work. Without his help, the equations upon which this research is based could not have been derived. I would like to thank Dr. Kenneth Chase for his expertise in uncertainty analysis, particularly in the field of tolerance analysis. His previous work has been a great reference and source of ideas for this research. I would like to thank Brad Larson for his work in compositional system models, for his advice, and for his patience and willingness in answering my incessant questions. I would like to thank BYU’s Department of Mechanical Engineering for its financial support of this research. Lastly, I would like to thank my parents for teaching me the value of hard work, my Heavenly Father for inspiring me at every step in this work, and my wife, Lacee, for her never-ending patience, support and love. TABLE OF CONTENTS LIST OF TABLES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi LIST OF FIGURES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii NOMENCLATURE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . x Chapter 1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 Chapter 2 Literature Survey . . . . . . . . . . . . 2.1 Error Propagation via Taylor Series Expansion 2.2 Non-Deterministic Analysis via Brute Force . 2.3 Deterministic Error by Model Composition . . 2.4 Error Budgets . . . . . . . . . . . . . . . . . . 2.5 Univariate Dimension Reduction . . . . . . . . 2.6 Interval Analysis Methods . . . . . . . . . . . 2.7 Bayesian Inference . . . . . . . . . . . . . . . 2.8 Response Surface Methodologies . . . . . . . . 2.9 Anti-Optimization Techniques . . . . . . . . . . . . . . . . . . . 3 3 4 5 5 6 6 6 6 7 Chapter 3 Propagation of Variance . . . . . . . . . . . . . . . . . . . . . . . . . 3.1 Fundamental Concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.1.1 Mathematical Expectation . . . . . . . . . . . . . . . . . . . . . . . . 3.1.2 Central Moments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2 Propagation of Variance Using First Order Taylor Series . . . . . . . . . . . 3.2.1 First-Order Formula Derivation . . . . . . . . . . . . . . . . . . . . . 3.2.2 First-Order Accuracy . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3 Propagation of Variance Using Higher-Order Taylor Series . . . . . . . . . . 3.3.1 Second-Order Formula Derivation . . . . . . . . . . . . . . . . . . . . 3.3.2 Second-Order Accuracy . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4 Taylor Series Error Propagation Assumptions and Limitations . . . . . . . . 3.5 Higher-Order Accuracy With First- or Second-Order Cost . . . . . . . . . . 3.5.1 Propagation of Variance Using Higher-Order Terms . . . . . . . . . . 3.5.2 Reducing Truncation Error . . . . . . . . . . . . . . . . . . . . . . . . 3.5.3 Correction Factors for Other Nonlinear Functions . . . . . . . . . . . 3.5.4 Model Composition . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.6 Example: Compositional System Model of a Dual Propeller, Three Degreeof-Freedom Helicopter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.7 Example: Kinematic Model of a Mechanism that Simulates Flapping Flight Wing Motion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.8 Comments on Variance Propagation . . . . . . . . . . . . . . . . . . . . . . . 9 9 9 10 11 11 13 14 15 16 17 18 18 20 22 24 Chapter 4 Higher-Order Statistics . . . . . . . . . . . . . . . . . . . . . . . . . 33 iv . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 28 30 4.1 4.2 4.3 4.4 4.5 4.6 Motivation for Propagating Higher-Order Statistics . . . . . . . . . . . . . . Propagation of Skewness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.1 Definition of Skewness . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.2 Skewness Propagation Formula Derivation . . . . . . . . . . . . . . . 4.2.3 Skewness Propagation Assumptions and Limitations . . . . . . . . . . 4.2.4 Proof: Nonlinear Functions Produce Non-Gaussian Output Distributions Propagation of Kurtosis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3.1 Definition of Kurtosis . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3.2 Definition of Excess Kurtosis . . . . . . . . . . . . . . . . . . . . . . . 4.3.3 Kurtosis Propagation Formula Derivation . . . . . . . . . . . . . . . . Example: Solar-Powered Unmanned Aerial Vehicle Propeller Thrust . . . . . 4.4.1 The Analytical Model of Thrust . . . . . . . . . . . . . . . . . . . . . 4.4.2 Statistical Component Model Output Prediction . . . . . . . . . . . . Example: Flat Rolling Metalworking Process . . . . . . . . . . . . . . . . . . 4.5.1 The Analytical Model . . . . . . . . . . . . . . . . . . . . . . . . . . 4.5.2 Statistical System Model Output Prediction . . . . . . . . . . . . . . 4.5.3 Ramifications of Neglecting Higher-Order Statistics . . . . . . . . . . Comments on Propagation of Higher-Order Statistics . . . . . . . . . . . . . 33 35 35 36 37 38 39 39 39 40 41 42 44 45 47 49 50 51 Chapter 5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53 REFERENCES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 Appendix AStep-by-Step Solution to Dual-Propeller, Three Degree-of-Freedom Helicopter Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 A.1 Variance Propagation Through Component 1 . . . . . . . . . . . . . . . . . . 60 A.2 Variance Propagation Through Component 2 . . . . . . . . . . . . . . . . . . 60 A.3 Variance Propagation Through Component 3 . . . . . . . . . . . . . . . . . . 62 A.4 Variance Propagation Through System-Level Compositional Model . . . . . 63 v LIST OF TABLES 3.1 3.2 3.3 3.4 Partial Derivatives Required for Taylor Series Variance Propagation . . . Truncation Error Correction Factors for Common Nonlinear Functions . Design Parameters for Dual-Propeller, 3-DOF Helicopter . . . . . . . . . Mean Values and Standard Deviations of the 16 Model Input Parameters . . . . . . . . 19 23 26 29 4.1 4.2 4.3 4.4 4.5 4.6 Comparison of Positive and Negative Skew . . . . . . . . . . . . . . . . . Statistical Properties of the Angular Velocity Distribution of a Propeller Central Moments of the Distribution of Angular Velocity, ω . . . . . . . . Friction Conditions in Mechanical Devices and Metalworking Processes . Statistical Properties of the Coefficient of Friction Distribution . . . . . . Central Moments of the Distribution of the Coefficient of Friction, µf . . . . . . . . . . . . . . . 35 43 44 47 48 49 A.1 Design Parameters for Dual-Propeller, 3-DOF Helicopter . . . . . . . . . . . 59 vi vii LIST OF FIGURES 3.1 3.2 3.3 3.4 3.5 3.6 3.7 3.8 3.9 3.10 3.11 3.12 3.13 4.1 4.2 4.3 4.4 4.5 4.6 4.7 4.8 4.9 Relative Error in Output Variance Using a First-Order Taylor Series Expansion for the Function y = 1000 sin(x) . . . . . . . . . . . . . . . . . . . . . . Relative Error in Estimate of Variance Propagation Using a Second-Order Taylor Series Expansion for the Function y = 1000 sin(x) . . . . . . . . . . . Relative Error in Variance Propagation Using Taylor Series Approximations Relative Computational Cost of Taylor Series Approximations, as Determined by MATLAB Execution Time . . . . . . . . . . . . . . . . . . . . . . . . . . Relationships Between the Second-Order Bias and both Input Standard Deviation σx and Input Variance σx2 for a sin Function . . . . . . . . . . . . . . Relative Error in Estimations of Variance Propagation . . . . . . . . . . . . Mean Relative Error in Variance Estimations . . . . . . . . . . . . . . . . . . Dual-Propeller, Three Degree-of-Freedom Helicopter Used in the Design of Unmanned Aerial Vehicles . . . . . . . . . . . . . . . . . . . . . . . . . . . . Relative Error in Estimations of Variance Propagation . . . . . . . . . . . . Mean Relative Errors that Indicate the Compositional Model with a Correction Factor Predicts Output Variance with About Half the Error as the First-Order Approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . Mechanism Used by the BYU Flapping Flight Research Team to Simulate Three-Degree-of-Freedom Motion of a Flapping Wing . . . . . . . . . . . . . Computational Time to Predict Output Distributions Using Various Error Propagation Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Relative Error in Predictions of Output Variance Obtained from Various Orders of a Taylor Series . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Predicted Gaussian Output Distribution Obtained from Propagating Mean and Variance Only (a) Compared with Actual System Output (b) . . . . . . Predicted Non-Gaussian Output Distribution Obtained from Propagating Mean, Variance, Skewness, and Kurtosis (a) Compared with Actual System Output (b) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Examples of Negative (Left) and Positive (Right) Skewness . . . . . . . . . . Excess Kurtosis of Various Common Statistical Distributions . . . . . . . . . Power and Propulsion System of a Solar-Powered Unmanned Aerial Vehicle (UAV) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Compositional System Model of a Solar-Powered Unmanned Aerial Vehicle (UAV) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Distribution of the Angular Velocities of a Propeller in a Solar-Powered UAV Predicted Gaussian Output Distribution of Thrust Obtained by Propagating a Mean and Variance Only (a) Compared with Actual Thrust Measurements (b) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Predicted Non-Gaussian Output Distribtuion of Thrust Obtained by Propagating a Mean, Variance, Skewness, and Kurtosis (a) Compared with Actual Thrust Measurements (b) . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii 14 17 19 20 21 22 22 25 27 27 28 29 30 34 34 36 40 42 43 44 45 46 4.10 Flat Rolling Manufacturing Process Whereby Plates or Sheets of Metal are Made—Material is Drawn Between Two Rollers, which Reduces the Material’s Thickness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11 Distribution of the Coefficient of Friction in a Flat Rolling Metalworking Process 4.12 Predicted Gaussian Output Distribution Obtained from Propagating a Mean and Variance Only (a) Compared with Actual System Output (b) . . . . . . 4.13 Predicted Non-Gaussian Output Distribution Obtained from Propagating a Mean, Variance, Skewness, and Kurtosis (a) Compared with Actual System Output (b) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix 46 48 50 50 NOMENCLATURE N n x xi xj x̄ σx σx2 σx2i xj γ1 γ2 β2 κi µk µk,i ∂1 ∂2 y ȳ σy σy2 e Ck E[ ] Number of data points in a distribution Number of inputs to a system model An input distribution to a system model The i-th system input distribution The j-th element of a system input distribution The mean value of input x Standard deviation in the input distribution x Variance in the input distribution x Covariance in between input distributions xi and Xj Skewness in a distribution Kurtosis in a distribution Excess kurtosis in a distribution The i-th cumulant of a distribution The k-th central moment of a distribution The k-th central moment of the i-th input distribution First partial derivative of the system model, evaluated at the mean Second partial derivative of the system model, evaluated at the mean System output Mean system output value Standard deviation in output distribution y Variance in output distribution y Relative error in a model’s prediction compared with actual output Correction factor The k-th component of a compositional system model indicates the expectation operator performed on [ ] Example: Dual Propeller, Three Degree-of-Freedom Helicopter θ Helicopter pitch angle (rad) θ̈ Angular acceleration in pitch (rad/s2 ) m1 Mass of helicopter motors, propellers, and circuitry (kg) m2 Mass of helicopter counterbalance weight (kg) Lk Length of helicopter segment k (m) Jy Mass moment of inertia about the pitch axis of rotation (kg-m2 ) uL Motor command given to left motor uR Motor command given to right motor km Motor calibration constant Example: Kinematic Model of Flapping Flight Wing Motion t Simulation time (s) ω Flapping frequency (Hz) φ Positional angle (deg) θ Elevation angle (deg) α Feathering or attack angle (deg) x Axx Bxx Fourier series coefficients for cos terms Fourier series coefficients for sin terms Example: Solar-Powered Unmanned Aerial Vehicle Propeller Thrust T Thrust generated by the propeller (N) Ct Unitless calibration coefficient of thrust ρ Density of air (kg/m3 ) ω Propeller angular velocity (rad/s) D Propeller Diameter (m) Example: Flat Rolling Metalworking Process ∆H The change in material thickness with each pass (m) ∆Hmax The maximum change in material thickness attainable in a single pass (m) µf Coefficient of friction R Roller radius (m) xi CHAPTER 1. INTRODUCTION In the system design process, designers frequently experience significant uncertainty in predicting whether a proposed design will meet the design objectives. Design decisions often cannot be validated until a physical prototype is built and tested. If the proposed design proves to be faulty, it can lead to a very costly design iterations. The inability of the system designer to verify design decisions early in the design process and to determine if a proposed design will accomplish design objectives is a limiting obstacle in system design [1]. Consequently, system behavioral modeling is an important part of system design. If a system behavioral model could be obtained and its accuracy quantified, it would enable the designer to verify design decisions early in the design process. This would greatly reduce the risk of a creating a failed system design. Determining system model accuracy is difficult, especially when the system is still theoretical and actual system behavior is not known [2]. In many situations, a statistical distribution of possible outputs is more meaningful than a simple max/min error bound [3]. If the system input distributions are known, such an output distribution can be obtained for closed-form differentiable models using a Taylor series expansion to propagate input distributions through a system model. This is typically accomplished with the first-order Taylor series approximation, but there are two major limitations of using this method. First, the first-order Taylor series takes a derivative-based weighted sum of independent input variances to estimate the variance in the output distribution. Since all higherorder terms in the series are neglected, this predicted output variance may be wrong by one or more orders of magnitude for nonlinear functions [4]. While higher-order terms clearly improve the accuracy of the approximation, they also require greater computational cost. This thesis shows that the truncation error in lower-order approximations can be predicted 1 and accounted for. Consequently, fourth-order accuracy in error propagation can be obtained without significant additional computational cost. Second, system designers typically use the Taylor series to only propagate a mean and variance. All higher-order statistics are neglected. As all non-Gaussian distributions cannot be fully described with a mean and variance alone, the output distribution is typically assumed to be Gaussian. This often is an erroneous and costly assumption. This thesis proves that nonlinear functions do not produce Gaussian output distributions, even when inputs are Gaussian. A second-order Taylor series expansion can be used to propagate higherorder statistical properties through a system model without incurring significant additional computational cost. This thesis explores error propagation through nonlinear system models using a Taylor series expansion. Chapter 2 provides an overview of nine common methods of uncertainty analysis. Chapter 3 shows how high accuracy in variance propagation can be obtained with low-order computational cost. Chapter 4 presents a method for propagating the higher-order statistical properties skewness and kurtosis through a closed-form system model. Chapter 5 concludes this thesis and suggests areas for future research. 2 CHAPTER 2. LITERATURE SURVEY Many methods are currently in use or being researched that quantify the accuracy of a system model by propagating error through the system. These methods include the following: • Error propagation via Taylor series expansion • Non-deterministic analysis via brute force (Monte Carlo, Latin hypercube, etc.) • Deterministic model composition • Error budgets • Univariate dimension reduction • Interval analysis • Bayesian inference • Response surface methodologies • Anti-optimizations (sub-optimizations) While this thesis focuses on error propagation via Taylor series expansion, the other uncertainly analysis techniques listed above are also briefly described in the remainder of this chapter. 2.1 Error Propagation via Taylor Series Expansion A Taylor series can be used to propagate variation in system inputs through a system model to produce an estimate of the variation in system outputs. This derivatives-based method can be very simple and fast, and is the method most-often cited in literature [5, 6]. 3 Nevertheless, it may produce results that are inaccurate by one or more orders of magnitude [4], particularly for nonlinear systems. This inaccuracy is a result of linearizing the system model, as the higher-order derivatives are truncated from the Taylor series. This linearization also means accuracy decreases as input variances increase. Typically, this method is used only to propagate variance. Data is typically assumed to be independent and Gaussian, as variable interactions and higher-order statistics are ignored. 2.2 Non-Deterministic Analysis via Brute Force Uncertainty analysis methods are usually classified as either deterministic or non- deterministic, and non-deterministic analysis methods are further grouped into two main categories: 1) reliability-based design methods [7, 8, 9], and 2) robust-design-based methods [10, 11, 12, 13, 14, 15]. A deterministic model always produces the same output values from the same input values, but a non-deterministic model does not. Consequently, a non-deterministic model, executed repeatedly with the same input values, can produce a statistical output distribution instead of a single nominal value. However, due to the complexities of non-deterministic modeling, it is more common to represent uncertainty with probabilistic methods and propagate these uncertainties through a deterministic model [16]. This approach can also result in a statistical output distribution by executing the model repeatedly. This brute-force approach to non-deterministic uncertainty analysis commonly uses Monte Carlo, quasi Monte Carlo [17, 18], Monte Carlo hybrid [19], Latin hypercube, Latin supercube [20], or some other sampling technique. This technique does not need to assume a Gaussian output distribution. Consequently, an estimate of the fully-described output distribution can be obtained. However, this approach comes at great computational cost, which only grows exponentially with an increase in the number of statistically distributed inputs. Furthermore, the entire simulation must be executed again each time the model or any input value changes. This can be prohibitive in an iterative design process. 4 2.3 Deterministic Error by Model Composition Current research is being done [21] to deterministically propagate error through a compositional system model to obtain max/min error bounds. In order to do this, the system model error must be included in the model itself. Augmenting the compositional system model with component error models allows the component error interactions and propagations to be evaluated together with the model. System model accuracy can be determined by comparing the results of the regular system model with those from the erroraugmented system model [2]. Monte Carlo simulations or optimization routines can be executed to find the max/min error bounds of the system model. With this approach to error propagation, the accuracy of even complex systems and interactions can be obtained. Errors do not need to be independent. Component models do not have to be mathematical or closed-form functions, but rather can include dynamic models, nonlinear models, non-differentiable models, lookup-tables, software programs, digital logic, CAD models, finite-element models, and any other model capable of mapping inputs to outputs. Deterministic error analysis requires known component error models and relationships. In some situations, the max/min error bounds obtained from this method are so large that they are not helpful. In many practical applications, a statistical error distribution is more useful than a max/min error envelop [3]. 2.4 Error Budgets The method of error budgets involves propagating the error of each component through the system separately, and resolving each component’s error to the contribution it makes on the total system error [22, 23]. This is done by perturbing one error source at a time and observing the effect this has on the total system error. Consequently, this method requires either that component errors be independent or that a separate model showing component error interactions be developed, which typically is not done [24]. If the error sources are not actually independent, this method will not necessarily describe the full range of possible model error. 5 2.5 Univariate Dimension Reduction Univariate dimension reduction methods transform data from a high-dimensional space to a lower-dimensional space. In some situations, data analysis may even be more accurate in the reduced space than in the original space [25]. 2.6 Interval Analysis Methods Interval analysis methods bound rounding and measurement errors in mathematical computation. Arithmetic can then be performed using intervals instead of a single nominal value [26]. These techniques can be used to propagate error envelopes, or intervals, through a system model. These methods, however, are typically limited to basic arithmetic operations. Currently there are many software languages, libraries, compilers, and data types that implement interval arithmetic. These include XSC, Profil/BIAS, Boost, Gaol, Frink, and a MATLAB extension named Intlab. There is also a working group currently developing an IEEE Interval Standard (P1788). 2.7 Bayesian Inference Bayesian inference is a method of statistical inference whereby the probability that a hypothesis is true is inferred based on both observed evidence and the prior probability that the hypothesis was true [27]. It combines common-sense knowledge with observational evidence in an attempt to eliminate needless complexity in a model by declaring only meaningful relationships [28] and disregarding the influences of all other variables on system outputs. 2.8 Response Surface Methodologies Response surface methodology is commonly used in design of experiments. It is a modeling technique whereby an n-dimensional response surface showing the relationship between n-input variables is created, typically from using empirical data [29, 30]. A few common experimental design setups include full factorial, partial factorial, central composite, Plackett-Burman, Box-Behnken, and others. 6 2.9 Anti-Optimization Techniques Anti-optimization techniques allow the designer to find the worst-case scenario for a given problem. This results in a two-level optimization problem, where the uncertainty is anti-optimized on the lower level and the overall design is optimized on a higher level [31]. 7 8 CHAPTER 3. PROPAGATION OF VARIANCE This chapter demonstrates how a Taylor series is used to propagate variance by estimating a distribution’s central moments. The mathematical expectation operator and central moments are first discussed. The derivations of the first- and second-order Taylor series error propagation formulas are then presented along with an in-depth discussion of their accuracies. The assumptions and limitations of using a Taylor series to propagate error through a system model are then summarized. This chapter then demonstrates how fourth-order accuracy can be obtained in error propagation with only first- or second-order computational cost. This is accomplished by applying a correction factor to the lower-order estimate, which accounts for the higher-order truncation error in the Taylor series. These correction factors were determined empirically for several common nonlinear engineering functions. Lastly, this chapter demonstrates the effectiveness of this method using the model of a dual-propeller, three degree-of-freedom helicopter, and the kinematic system model of a flapping wing. 3.1 Fundamental Concepts Before continuing, it is essential to understand two statistical concepts that are funda- mental to this research. These are the mathematical expectation operator and the statistical central moment property, both of which are described in this section. 3.1.1 Mathematical Expectation The mathematical expectation operator, E[ ] denotes the calculation of a weighted average of all possible values, as shown in Eq. (3.1). The weights correspond to the probability of a particular value. When all possible values have an equal probability, the expected 9 value is equal to the arithmetic mean, which is also equal to the limit of the sample mean as the sample size increases to infinity. E[X] = x1 p1 + x2 p2 + ... + xn pn (3.1) where X is some population, xi are the possible values, pi are their respective probabilities, and all pi add to 1. Some common expectation properties are presented below: E[a] = a E[X + a] = E[X] + a E[X + Y ] = E[X] + E[Y ] E[a · X] = a · E[X] E[aX + b] = a · E[X] + b E[aX + bY ] = a · E[X] + b · E[Y ] where X and Y are statistical distributions and a and b are constants. 3.1.2 Central Moments A central moment is a statistical property commonly used in statistical analysis [32, 6, 33]. The k-th central moment of a distribution is defined in Eq. (3.2). µk = E (x − x̄)k = N 1 X (xj − x̄)k N j=1 (3.2) where x represents some distribution of N values, x̄ represents the input mean, and E is the mathematical expectation operator. The central moments of a population can easily be estimated using any appropriate population-sampling technique. 10 The zeroth central moment is always equal to one, the first central moment is always equal to zero [34], and the second central moment is equivalent to the variance. The third and fourth moments are used in the calculation of higher-order statistics, such as skewness and kurtosis. Propagation of these higher-order statistics is the subject of Chapter 4. 3.2 Propagation of Variance Using First Order Taylor Series Generally, a distribution of input values propagates through a given system to pro- duce a distribution of output values. While many sources of error might be present, including modeling error (unmodeled behavior, emergent behavior) and measurement error, this research focuses on the output distribution caused by variation in system model inputs only. The analytical formula most-often cited in literature that estimates this variance propagation is based on a first-order Taylor series expansion. It makes several important assumptions that are limiting in many practical situations. These assumptions and limitations cannot be fully understood (much less overcome) without understanding the derivation of this first-order variance propagation formula. The author of this thesis has found the complete derivation of this formula to be absent in textbooks and archival journal literature. Consequently, the derivation is presented in this section in order to more fully explain the assumptions, limitations, and accuracy of this error propagation technique. The assumptions and limitations mentioned are discussed in Section 3.4. 3.2.1 First-Order Formula Derivation Let y be some function of n inputs xi . The first-order Taylor series approximation expanded about the input means x̄i is shown in Eq. (3.3) [35]. n X ∂f y ≈ f (x̄1 , ..., x̄n ) + (xi − x̄i ) ∂xi i=1 (3.3) where the partial derivatives are evaluated at the mean xi = x̄i . An approximation of the output mean ȳ is given in Eq. (3.4). 11 ȳ = E[y] " # n X ∂f ≈ E f (x̄1 , ..., x̄n ) + (xi − x̄i ) ∂x i i=1 n X ∂f µ1,i ≈ f (x̄1 , ..., x̄n ) + ∂xi i=1 ≈ f (x̄1 , ..., x̄n ) (3.4) where E is the expectation operator and µk,i is the k-th central moment for the i-th input, as given previously in Eq. (3.2). (Recall that the first central moment is equal to zero.) Subtracting Eq. (3.3) from Eq. (3.4) produces Eq. (3.5). n X ∂f y − ȳ ≈ (xi − x̄i ) ∂xi i=1 (3.5) Squaring and taking the expectation of Eq. (3.5) produces Eq. (3.6). " # 2 n n X X ∂f ∂f ∂f 2 σ E[(y − ȳ)2 ] ≈ σx2i + 2 ∂xi ∂xi ∂xj xi xj i=1 j=i+1 = σy2 (3.6) where σy2 and σx2 are the variances in y and x, respectively. Recall that variance σ 2 is the second central moment, which is defined in Eq. (3.7). σx2 = µ2 = E[(x − x̄)2 ] N 1 X = (xj − x̄)2 N j=1 (3.7) The second term in Eq. (3.6) is the covariance term, where σx2i xj is the covariance between inputs xi and xj . Covariance is defined in Eq. (3.8). 12 σx2i xj = E [(xi − x̄i )(xj − x̄j )] (3.8) When inputs are independent, the covariance term is equal to zero, and Eq. (3.6) reduces to Eq. (3.9). σy2 ≈ 2 n X ∂f i=1 ∂xi σx2i (3.9) This simplifying assumption of independence is typically made, both in literature and in practice. Consequently, Eq. (3.9) is the formula typically given for statistical error propagation through an analytical system model [36, 37, 38, 4]. 3.2.2 First-Order Accuracy Clearly, Eq. (3.9) is an approximation only and can be wrong by one or more orders of magnitude. This is especially evident when dealing with nonlinear functions [4]. For example, let y be modeled by the function y = 1000 sin(x). An estimation of the output variance σy2 obtained from Eq. (3.9) is given in Eq. (3.10). 2 σy,1st ≈ 106 cos2 (x̄)σx2 (3.10) Throughout this entire thesis, “actual” output variance was determined using a Monte Carlo simulation. While convergence in statistical properties was obtained for every model in this thesis after 1 million executions, each system’s “actual” output is the result of 10 million executions. The relative error of Eq. (3.10) can be calculated using Eq. (3.11) and the result is plotted as a function of input mean x̄ in Figure 3.1. = 2 |σy, 2 predicted − σy,M onte Carlo | 2 σy,M onte Carlo (3.11) As illustrated in Figure 3.1, the first-order approximation of variance propagation is fairly accurate for most values of x̄. However, for certain input values, the approximation can be wrong by one or more orders of magnitude, as indicated by the 100% jump in relative 13 Relative Error in 1st Order Variance Prediction Relative Error (%) 100 80 60 40 20 0 0 0.5 1 1.5 2 Input Mean 2.5 3 Figure 3.1: Relative Error in Output Variance Using a First-Order Taylor Series Expansion for the Function y = 1000 sin(x) error at x̄ = π2 . This spike in relative error occurs because the sin function is nonlinear and the higher-order terms in the Taylor series used to derive Eq. (3.9) were neglected. 3.3 Propagation of Variance Using Higher-Order Taylor Series As shown in the preceding section, Eq. (3.9) is based on a first-order Taylor series. For nonlinear and higher-order polynomial functions, Taylor series truncation error becomes significant and Eq. (3.9) can become extremely inaccurate (i.e., wrong by one or more orders of magnitude [4]). As expected, the accuracy of this estimate of variance propagation through a system can be improved by including higher-order terms in the Taylor series. In situations where increased accuracy is required, a second-order Taylor series is sometimes used to propagate statistical error. Similar to the first-order approximation, some simplifying assumptions are commonly made in the derivation of the second-order approximation, which are presented in Section 3.4. This section presents the derivation of the second-order error propagation formula and then discusses the accuracy of this formula. 14 3.3.1 Second-Order Formula Derivation For the sake of brevity, the second-order derivation is presented in this section for a monovariable function, y = f (x). Extending this derivation to multivariate functions is trivial, as it follows the same derivation steps. The second-order Taylor series taken about the input mean x̄ is given in Eq. (3.12), where the partial derivatives are again evaluated at the mean, x = x̄. y ≈ f (x̄) + ∂f 1 ∂ 2f (x − x̄) + (x − x̄)2 ∂x 2 ∂x2 (3.12) The second-order approximation of the output mean ȳ is given in Eq. (3.13). ȳ = E[y] 1 ∂ 2f ∂f 2 (x − x̄) + (x − x̄) ≈ E f (x̄) + ∂x 2 ∂x2 1 ∂ 2f ≈ f (x̄) + µ2 2 ∂x2 (3.13) Subtracting Eq. (3.13) from Eq. (3.12) gives Eq. (3.14). y − ȳ ≈ ∂f 1 ∂ 2f 1 ∂ 2f 2 (x − x̄) + (x − x̄) − µ2 ∂x 2 ∂x2 2 ∂x2 (3.14) Squaring and taking the expectation of Eq. (3.14) produces Eq. (3.15). E (y − ȳ)2 ≈ µ2 ∂f ∂x 2 ∂f ∂ 2 f 1 2 + µ3 + µ − µ 4 2 ∂x ∂x2 4 ≈ σy2 ∂ 2f ∂x2 2 (3.15) If x is Gaussian, all odd moments (µk where k is odd) are zero and Eq. (3.15) reduces to Eq. (3.16). 15 σy2 ∂f ∂x 2 ∂f ∂x 2 ≈ ≈ 1 µ2 + 4 ∂ 2f ∂x2 2 1 + 4 ∂ 2f ∂x2 2 σx2 (µ4 − µ22 ) (µ4 − σx4 ) (3.16) Furthermore, if x is Gaussian, the substitution µ4 ≈ 3σ 4 can be made [39], which eliminates the need to know the input’s higher-order moments. This substitution is made in Eq. (3.17). σy2 ≈ ∂f ∂x 2 σx2 1 + 2 ∂ 2f ∂x2 2 σx4 (3.17) If y is a function of multiple independent inputs, the generalized form of Eq. (3.17) is given in Eq. (3.18). σy2 ≈ 2 n X ∂y i=1 ∂xi n σx2i n 1 XX + 2 j=1 i=1 ∂ 2y ∂xi ∂xj 2 σx2i σx2j (3.18) Equation (3.18) is the second-order formula most often cited in literature [40, 36] for analytical statistical error propagation. Note that the covariance terms in Eq. (3.18) have been neglected. 3.3.2 Second-Order Accuracy Continuing with the same function y = 1000 sin(x), the relative errors obtained from the second-order approximations in Eq. (3.15) (full approximation) and Eq. (3.16) (Gaussian) are plotted as a function of input mean x̄ in Figure 3.2. Figure 3.2 illustrates the noise introduced by assuming x is Gaussian when in reality the distribution is never perfect. Nevertheless, in both cases the second-order approximation successfully filters the large spikes in relative error present in the first-order approximation. However, the second-order approximation still overestimates the actual variance propagation, which could degrade system performance and lead to failure or infeasibility [41]. This bias (about 4%, in this case) is a result of truncating the higher-order terms in the Taylor series. 16 Relative Error in 2nd Order Approximation Relative Error (%) 4.5 4 Gauss. Full 3.5 0 0.5 1 1.5 2 Input Mean 2.5 3 Figure 3.2: Relative Error in Estimate of Variance Propagation Using a Second-Order Taylor Series Expansion for the Function y = 1000 sin(x) 3.4 Taylor Series Error Propagation Assumptions and Limitations Though common in engineering literature and academia, Eqs. (3.9) and (3.18) have many significant limitations. Often, designers use these equations without knowing or understanding these limitations. The following list summarizes these assumptions and limitations: 1. Only variance is propagated and higher-order statistics are neglected. Consequently, the output distribution is generally assumed to be a Gaussian distribution, which is not the case in many practical applications. 2. The system model y must be representable as a closed-form, differentiable, mathematical equation. 3. Taking the Taylor series expansion about a single point (x̄) causes the approximation to be of local validity only [36, 19]. Consequently, the accuracy of the approximation generally decreases with an increase in the input variance σx2 . 4. The approximation is generally more accurate for linear models. 5. All inputs xi are assumed be Gaussian. Consequently, only means and variances are used to fully describe the input distributions. When inputs are not Gaussian, the non-Gaussian terms (i.e., odd moments) cannot be neglected. 17 6. The input means and standard deviations must be known. 7. All inputs xi are assumed to be independent. When inputs are not independent, the covariance terms cannot be neglected [37, 42, 4]. The first assumption listed above is addressed in detail in Chapter 4. 3.5 Higher-Order Accuracy With First- or Second-Order Cost While an improvement over the first-order formula, the second-order equation given by Eq. (3.18) is far from perfect. This section shows that accuracy can be improved by adding higher- and higher-order terms. However, the resulting computational cost quickly becomes prohibitive. This is a common problem in engineering analysis and design [43]. This section then presents a solution to that problem. This is accomplished by predicting the higher-order truncation error and applying a resultant correction factor to the lower-order approximation. This can yield fourth-order accuracy in the estimation of error propagation with first- or second-order computational cost. These correction factors were empirically determined for trigonometric, logarithmic, and exponential functions. 3.5.1 Propagation of Variance Using Higher-Order Terms As expected, adding higher-order terms reduces the second-order bias. This is shown in Figure 3.3. With an infinite number of terms, the Taylor series approximation eventually converges to zero error. However, computational cost grows exponentially as higher-order terms are included. This growth in computational cost is accelerated at an exponential rate with an increase in the number of system inputs, as shown in Table 3.1. Furthermore, higher-order terms also require the calculation of higher-order moments and covariance terms for the system inputs. This exponential growth in cost causes higher-order terms to quickly become prohibitively expensive for complex systems. For the purposes of this research, the computational costs of propagating error were determined using the MATLAB execution time. Figure 3.4 shows the relative computational 18 Relative Error Using Higher−Order Taylor Series 6 O1 Relative Error (%) 5 O2 4 O3 3 O4 2 1 0 0 0.5 1 1.5 2 Input Mean 2.5 3 Figure 3.3: Relative Error in Variance Propagation Using Taylor Series Approximations Table 3.1: Partial Derivatives Required for Taylor Series Variance Propagation Inputs 2 4 6 8 10 O1 2 4 6 8 10 O2 O3 O4 O5 O6 3 6 10 15 21 10 20 35 56 84 21 56 126 252 462 36 120 330 792 1716 55 220 715 2002 5005 O7 28 120 792 3138 11146 cost associated with including higher-order terms in the calculation of variance propagation. (The second-order approximation with a correction factor seen at the end of Figure 3.4 is presented and discussed in Section 3.5.2.) Furthermore, computational cost grows exponentially with an increase in the number of system inputs and with an increase in model complexity. As this high cost can be prohibitive, higher-order terms are usually not considered. The author of this thesis has shown that higher-order truncation error can be predicted. Consequently, a correction factor can be applied to the lower-order formula, yielding higher-order accuracy without significantly increasing computational cost. 19 Computation Cost Computation Cost of Higher−Order Terms 1st Full 2nd Full 3rd Full 4th Full 2nd Corrected Figure 3.4: Relative Computational Cost of Taylor Series Approximations, as Determined by MATLAB Execution Time 3.5.2 Reducing Truncation Error As Figure 3.2 illustrates, the truncation error in the second-order approximation for a sin function is essentially a constant bias for all x̄. Furthermore, this bias has a linear relationship to the input variance, σx2 , as shown in Figure 3.5. Consequently, the second-order truncation error can be easily be estimated empirically. A correction factor e corresponding to this truncation error can be calculated using Eq. (3.19). e= 1 1 + 1.022σx2 (3.19) This correction factor can then be applied to the second-order approximation of variance propagation, as shown in Eq. (3.20), which reduces the higher-order Taylor series truncation error. 2 σy,CF ≈ σy2 e (3.20) By way of example, consider the same function y = 1000 sin(x). The corrected second-order approximation produces an estimate of error propagation with fourth-order accuracy, but with only second-order computational cost. Figure 3.4 (already displayed in 20 2nd Order Mean % Err. 10 Mean % Error 8 6 4 2 0 0 0.05 0.1 0.15 0.2 Input Standard Deviation (σx) 0.25 0.3 2nd Order Mean % Err. 10 Mean % Error 8 6 4 2 0 0 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 2 Input Variance (σx) Figure 3.5: Relationships Between the Second-Order Bias and both Input Standard Deviation σx and Input Variance σx2 for a sin Function Section 3.5.1) compares the computational costs of the uncorrected Taylor series approximations with the corrected second-order approximation. Figure 3.6 compares the relative errors in variance estimations. Note that the fourthorder and second-order corrected approximations are almost identical in Figure 3.6. Figure 3.7 compares the mean relative error averaged over the range 0 ≤ x̄ ≤ π, showing that fourth-order accuracy can be obtained by adding a correction factor to the second-order formula for error propagation. 21 Relative Error in Variance Estimation Relative Error (%) 5 4 O2 3 O4 2 CF 1 0 0 0.5 1 1.5 2 Input Mean 2.5 3 Figure 3.6: Relative Error in Estimations of Variance Propagation Accuracy Summary of Variance Propagation Formulas Mean Relative Error (%) 5 2nd Order 4th Order 2nd Corrected 4 3 2 1 0 Estimation Method Figure 3.7: Mean Relative Error in Variance Estimations 3.5.3 Correction Factors for Other Nonlinear Functions Table 3.2 gives the correction factors for some common nonlinear functions. Note that the cyclical nature of trigonometric functions and derivatives require a second-order approximation before the higher-order truncation error can easily be determined, but the exponential and logarithmic functions only require a first-order calculation. The correction factors for y = ln(x), y = exp(x), and y = bx are given in Eqs. (3.21), (3.22), and (3.23), respectively. Correction factor for y = ln(x), where r = ln 22 x̄ σx : Table 3.2: Truncation Error Correction Factors for Common Nonlinear Functions Func. Ord. y = sin(x) 2nd y = cos(x) 2nd y = ln(x) 1st y = exp(x) 1st y = bx 1st e= Correction Factor (e) Eq. (3.19) Eq. (3.19) Eq. (3.21) Eq. (3.22) Eq. (3.23) exp (−1.9772r + 0.9128) + 1 if r >= 0 (3.21) −3.88r3 − 4.9835r2 − 1.5704r + 1.3302 if r < 0 Correction factor for y = ex , where e is constrained to a maximum value of 2 and a minimum value of 1: e = max 1, min 2, 0.3375σx2 + 0.4937σx + 0.959 (3.22) Correction factor for y = bx , where e is constrained to a maximum value of 2 and a minimum value of 1: h i max 1, min 2, X · (Z1 · B)T + 1 if b < 1 e= h i max 1, min 2, X · (Z2 · B)T + 1 if b > 1 where X= h 1 B= h σx2 σx 1 b 23 σx3 2 b b 3 σx4 iT i (3.23) 0.224 −0.992 Z1 = −0.037 0 0.012 0 0.076 1.544e4 0 7.803 −2.061e4 7.135 0 0 23.556 −8.612 1.182 1 Z2 = 1000 0.101 −0.002 3.5.4 −2.570e3 −897.1 0.018 0 0 −38.093 61.532 0.006 0.084 986.2 −0.002 −6.891 54.93 −240.9 0.089 −0.276 −1.743 0.005 −0.008 26.52 11.06 Model Composition It should be noted that the correction factors given in Table 3.2 are only pertinent to a particular function. If a system model contains this function along with other operators, the system should be decomposed into components (with sin(x) being a single component, for example). The error should then be propagated through each component individually. The variances in each component’s output can then be propagated through the rest of the system model. This process of model decomposition is demonstrated in Section 3.6 with the example of a dual-propeller, three degree-of-freedom helicopter. 3.6 Example: Compositional System Model of a Dual Propeller, Three Degreeof-Freedom Helicopter Consider the dual-propeller, three degree-of-freedom helicopter shown in Figure 3.8. This helicopter is a useful tool for the design of unmanned aerial vehicles (UAVs). Among other benefits, it allows the system designer to 1) test pitch, roll, and yaw control systems without endangering actual aircraft, 2) observe the effects of changing the payload and component weights and positions, and 3) calibrate the conversion factor (given as km below) that converts a motor’s throttle command into units of thrust. A UAV’s position and orientation are the inputs to its control system. An efficient prediction of the error in the UAV’s position allows the designer to easily create a robust control system capable of accomplishing mission objectives. 24 Figure 3.8: Dual-Propeller, Three Degree-of-Freedom Helicopter Used in the Design of Unmanned Aerial Vehicles The system model for the longitudinal (pitch) acceleration as a function of the system’s design parameters and inputs is given in Eq. (3.24). θ̈ = (m2 L2 − m1 L1 )g L1 km (uL + uR ) cos θ + 2 2 m1 L1 + m2 L2 + Jy m1 L21 + m2 L22 + Jy (3.24) where the masses mi and lengths Li are indicated in Figure 3.8, Jy is the mass moment of inertia about the pitch axis of rotation, g is the acceleration of gravity, uL and uR are the throttle commands (0-100) given to the left and right motors, respectively, and km is a motor calibration constant that converts throttle commands to units of thrust. For this demonstration, the assumed mean and standard deviation of the dependent and independent system design parameters are given in Tab. 3.3. Three system inputs are also required: uL and uR , which were each fixed at 45, and the pitch angle θ. The mean value of θ was varied from −45◦ to 45◦ , with a standard deviation of 1◦ . The first-order approximation in Eq. (3.9) can be used to propagate error through this system (see Figures 3.9-3.10). However, the nonlinearity of the cos function causes significant error in the results. Consequently, the model should be decomposed into its components, as shown in Eq. (3.25). Error can then be propagated through each component. Once the variance in each component is obtained, each component can then be treated as a variable 25 Table 3.3: Design Parameters for Dual-Propeller, 3-DOF Helicopter Design Parameter Mean m1 0.891 (kg) m2 1.000 (kg) L1 0.850 (m) L2 0.3048 (m) Jy 0.0014 (kg · m2 ) km 0.0546 Std. Dev. 10−4 (kg) 10−4 (kg) 10−4 (m) 10−4 (m) 10−5 (kg · m2 ) 10−5 and its variance propagated through the system-level compositional model. A full step-bystep solution is presented in Appendix A. θ̈ = C1 C2 + C3 (m2 L2 − m1 L1 )g C1 = m1 L21 + m2 L22 + Jy C2 = cos θ L1 km (uL + uR ) C3 = m1 L21 + m2 L22 + Jy (3.25) The first-order Taylor series approximation given in Eq. (3.9) is used to propagate error from the input parameters through components C1 and C3 . The second-order approximation with a correction factor is used to propagate error through the nonlinear trigonometric component C2 . Figure 3.9 compares the results of this compositional model with a correction factor applied with the results of the first-order approximation from Eq. (3.9) applied to the full system model in Eq. (3.24). These results are shown for −45◦ ≤ θ̄ ≤ 45◦ . Figure 3.10 shows the relative error for each of these methods, averaged across the entire input domain −45◦ ≤ θ̄ ≤ 45◦ . The computational cost required to determine the output variance using the compositional system model was approximately three times less than the cost using the first-order approximation. Thus, twice the accuracy was obtained with one-third of the computational 26 Relative Error Relative Error (%) 40 30 Compositional 1st Order 20 10 0 −40 −20 0 20 θ Mean (deg) 40 Figure 3.9: Relative Error in Estimations of Variance Propagation Mean Relative Error Mean Relative Error (%) 2 1.5 1 0.5 0 1st Order Approximation Compositional Model Estimation Method Figure 3.10: Mean Relative Errors that Indicate the Compositional Model with a Correction Factor Predicts Output Variance with About Half the Error as the First-Order Approximation cost. This efficiency in error propagation allows the system designer to more easily create a robust system capable of accomplishing mission objectives. 27 3.7 Example: Kinematic Model of a Mechanism that Simulates Flapping Flight Wing Motion Consider the flapping flight wing mechanism shown in Figure 3.11. The kinematic model used in the design and optimization of this mechanism is the Fourier series in Eq. (3.26) [44, 45]. Figure 3.11: Mechanism Used by the BYU Flapping Flight Research Team to Simulate Three-Degree-of-Freedom Motion of a Flapping Wing φ(t) A B 2 φn X φn θ(t) = Aθn cos(nωt) + Bθn sin(nωt) n=0 α(t) Aαn Bαn (3.26) where φ is the positional angle (deg), θ is the elevation angle (deg), α is the feathering or attack angle (deg), the As and Bs are Fourier series coefficients (deg), ω is the flapping frequency (Hz), and t is time (s). This three-output system model has 16 Gaussian inputs (time does not vary), which are statistically described in Table 3.4 [45]. Various orders of a Taylor series were used to estimate the variance in the three output wing angles based on the variance in these system inputs. The system model was then decomposed (as demonstrated in the previous example) and the trigonometric correction 28 Table 3.4: Mean Values and Standard Deviations of the 16 Model Input Parameters Aφ0 Aφ1 Aφ2 Aθ0 Aθ1 Aθ2 Aα0 Aα1 Aα2 ω Mean -20 -4 8 0 43 17 12 0 -2 0.298 StdDev 0.1 1.5 3.0 0.1 0.75 0.5 4.0 0.1 0.75 0.1 Bφ1 Bφ2 Bθ1 Bθ2 Bα1 Bα2 Mean 44 33 0 0 50 0 StdDev 1.5 8.0 0.1 0.1 0.1 0.1 factors given in Table 3.2 were then applied. Figure 3.12 shows the computational cost of error propagation (in minutes) using these different methods. These costs are summed over the time interval 0s≤t≤5s at time steps of 0.001s for all three output angles. The cost of calculating input moments and input covariances from 10k input samples is also included. Computational cost was determined using MATLAB execution time. Computational Time of Variance Propagation Methods Execution Time (min) 80 60 O1 O2 O3 O4 CF 40 20 0 Estimation Method Figure 3.12: Computational Time to Predict Output Distributions Using Various Error Propagation Methods 29 The fourth-order prediction took approximately 70 minutes to execute, where the corrected second-order approximation executed in approximately 4 minutes—a significant reduction in computational cost. The relative error in only one of these output angles, φ, is shown in Figure 3.13, as the other two output angles have similar results. The root-mean-square of the relative error over the time interval shown (0s≤t≤5s) for the second-order approximation is 40.97%, third-order is 11.18%, fourth-order is 1.32%, and a second-order with a correction factor is 1.96%. Relative Error in φ Relative Error (%) 150 O2 O3 O4 CF 100 50 0 0 1 2 3 4 5 Time (s) Figure 3.13: Relative Error in Predictions of Output Variance Obtained from Various Orders of a Taylor Series Figures 3.12 and 3.13 illustrate that the correction factors presented in this paper can achieve near fourth-order accuracy in error propagation through this model with near second-order computational cost. 3.8 Comments on Variance Propagation Using a first-order Taylor series to estimate error propagation through a closed-form model is common practice. Frequently, this approximation is used without a full appreciation of its limitations and the assumptions upon which it is based. This results in estimations that may be substantially inaccurate. Furthermore, the author has been unable to locate the full 30 derivation of this formula either in textbook or archival journal literature. Consequently, a useful contribution of the research presented in this chapter is the derivation and the discussion of the limitations and assumptions of the first- and second-order Taylor series approximations for variance propagation. Additionally, the novel contribution of this chapter is the introduction of generic correction factors that account for some of the Taylor series truncation error for common nonlinear functions encountered in engineering. These correction factors are predictable, easy to calculate, and don’t require significant computational cost. A system designer can use these correction factors to predict error propagation through trigonometric, logarithmic, and exponential functions with greater accuracy without greater computational cost. This enables the designer to better verify design decisions, which reduces the risk of developing a design that does not meet design objectives. Future research may focus on the development of predictable correction factors for other nonlinear models, such as differential equations and state-space models. The author believes this can be accomplished using the same methods used in this chapter. 31 32 CHAPTER 4. HIGHER-ORDER STATISTICS For closed-form, analytical models, a statistical error distribution is usually obtained by propagating variance from system inputs to system outputs using a Taylor series. This has already been demonstrated in Chapter 3. As noted in Section 3.4, the variance propagation formulas given in Eqs. (3.9) and (3.18) assume all inputs are Gaussian. Since all higher-order statistics (e.g., skewness, kurtosis, etc.) are ignored, outputs are also typically assumed to be Gaussian. This assumption is often erroneous and does not accurately reflect reality, as proved later in Section 4.2.4. In order to truly perform an accurate statistical error analysis, outputs cannot be assumed to be Gaussian. However, since all non-Gaussian distributions cannot be fully described by a mean and standard deviation alone, higher-order statistics (e.g., skewness and kurtosis) must also be used. Fortunately, this is relatively simple using the method and formulas presented in this chapter. This chapter first demonstrates the need for propagating higher-order statistics. A second-order Taylor series is then used to show how skewness and kurtosis can both be propagated through a system model. Having a mean, variance, skewness, and kurtosis, the system designer can then fully describe the output distribution. The benefits of obtaining a fully-described output distribution are demonstrated using two examples: the propeller component of a solar-powered unmanned aerial vehicle, and a flat rolling metalworking process. 4.1 Motivation for Propagating Higher-Order Statistics Consider the simple quadratic function, y = x2 . Assume the input x is a Gaussian distribution with a mean x̄ and a standard deviation σx both equal to 1. Equation (3.9) can be used to propagate this input distribution through the system model and predict the 33 Gaussian output distribution shown in Figure 4.1a. This predicted output is very different from the actual system output distribution, shown in Figure 4.1b. Actual Output Distribution 0.4 0.3 0.3 Prob. (%) Prob. (%) Output Predicted from Mean/Var. Only 0.4 0.2 0.1 0 0.2 0.1 −5 0 5 Output Value (y) 0 10 (a) Gaussian Prediction Using Mean and Variance Only −5 0 5 Output Value (y) 10 (b) Actual System Output Figure 4.1: Predicted Gaussian Output Distribution Obtained from Propagating Mean and Variance Only (a) Compared with Actual System Output (b) However, if skewness and kurtosis are also propagated through the system model, the predicted output distribution resembles actual system output much more closely [46]. This is illustrated in Figure 4.2. Actual Output Distribution 0.4 0.3 0.3 Prob. (%) Prob. (%) Output Predicted from Mean, Var., Skew., & Kurt. 0.4 0.2 0.1 0 0.2 0.1 −5 0 5 Output Value (y) 0 10 (a) Non-Gaussian Prediction Using Mean, Variance, Skewness, and Kurtosis −5 0 5 Output Value (y) 10 (b) Actual System Output Figure 4.2: Predicted Non-Gaussian Output Distribution Obtained from Propagating Mean, Variance, Skewness, and Kurtosis (a) Compared with Actual System Output (b) A proof that nonlinear systems produce non-Gaussian outputs even when inputs are Gaussian is presented later in Section 4.2.4. 34 4.2 Propagation of Skewness As previously noted, non-Gaussian distributions cannot be fully described with a only mean and standard deviation. Consequently, higher-order statistics, such as skewness and kurtosis, must also be used to describe non-Gaussian distributions. This section considers the definition of skewness and derives a formula for propagating skewness through an analytical system model. 4.2.1 Definition of Skewness The first-order statistic of a distribution is its mean, the second-order statistic is its standard deviation, and the third-order statistic is its skewness. Skewness is a measure of a distribution’s asymmetry. Skewness (denoted γ1 ) is defined in Eq. (4.1). " γ1 = E = x − x̄ σ 3 # µ3 σ3 (4.1) where E is the expectation operator, µ3 is the third central moment, σ is the standard deviation. Table 4.1 and Figure 4.3 illustrate some characteristics and terminology of positivelyand negatively-skewed distributions. A skewness of zero indicates a symmetric distribution. Table 4.1: Comparison of Positive and Negative Skew Sign Left/Right Mean vs. Median [47] Negative Left-skewed Mean is typically (though not always) less than the median Positive Right-skewed Mean is typically (though not always) greater than the median Skewness is an important defining characteristic of statistical distributions. A measure of skewness is required to fully describe any asymmetric distribution. Traditional uncertainty propagation, however, only propagates a mean and variance. With no skewness 35 Figure 4.3: Examples of Negative (Left) and Positive (Right) Skewness information available, skewness is neglected (assumed equal to zero) and a Gaussian distribution is assumed. 4.2.2 Skewness Propagation Formula Derivation Using a first-order Taylor series to propagate skewness through a system model results in an output skewness equal to the input skewness. It has already been demonstrated that this often does not reflect reality, even for simple nonlinear functions, and a proof is presented later in Section 4.2.4. Consequently, a second-order Taylor series will be used to derive a formula for skewness propagation. The second central moment of output y has already been given in Eq. (3.15). The third central moment is given by the dot-product in Eq. (4.2). 3 E (y − ȳ) ≈ µ3 ∂13 2 ∂1 ∂2 (µ4 − · 2 3 3 ∂1 ∂2 µ − 2 µ2 µ3 4 5 1 3 1 3 3 µ − 8 µ2 µ4 + 8 µ6 ∂2 4 2 3 2 µ22 ) (4.2) where µk is the k-th central moment of input x, and ∂1 and ∂2 respectively represent the partial derivatives ∂f ∂x and ∂2f , ∂x2 evaluated at the mean x = x̄. The third moment is a cubic function, and consequently it has four terms. Equation (4.2) has both first and second partial derivatives, because it is based on a second-order Taylor series. If a higher-order Taylor series were used, Eq. (4.2) would contain higher-order partial derivatives. 36 The second moment from Eq. (3.15) and the third moment from Eq. (4.2) can be used with the definition of skewness given by Eq. (4.1) to estimate the skewness in output y. This output skewness estimation is given in Eq. (4.3). γ1 E (y − ȳ)3 = 1.5 E (y − ȳ)2 µ3 ∂13 2 3 2 ∂1 ∂2 (µ4 − µ2 ) 2 · 2 3 3 ∂1 ∂2 µ − 2 µ2 µ3 4 5 1 3 3 1 3 µ − 8 µ2 µ4 + 8 µ6 ∂2 4 2 ≈ 1.5 µ2 ∂12 + µ3 ∂1 ∂2 + 41 (µ4 − µ22 ) ∂22 (4.3) Equation (4.3) estimates output skewness using the input central moments, µk . If input skewness, kurtosis, and higher-order statistics are known instead of input moments, these statistics can easily be substituted into Eq. (4.3) in place of these moments. For the sake of brevity, the skewness propagation formula has only been derived for mono-variate functions. However, this derivation can easily been extended to multivariate functions as desired. It should be noted that the most computationally expensive part to propagating skewness is calculating first and second derivatives. However, these have already been calculated in order to propagate variance if a second-order Taylor series was used, and consequently the additional cost to also propagate skewness is minimal. 4.2.3 Skewness Propagation Assumptions and Limitations The following five assumptions and limitations apply to the method just presented to propagate skewness: 1. Equation (4.3) is based on a second-order Taylor series. Consequently, it will predict output skewness perfectly for second-order (or lower) functions. Accuracy decreases with increasing non-linearity. 37 2. While not a problem for most analytical engineering models, it should be noted that this method of skewness propagation requires the system model to be represented as a closed-form, twice-differentiable, mathematical equation. This is due to the dependence of this method on a Taylor series. 3. The system model y must be representable as a closed-form, differentiable, mathematical equation. 4. Taking the Taylor series expansion about a single point (x̄) causes the approximation to be of local validity only [36, 19]. Consequently, the accuracy of the approximation generally decreases with an increase in the input moments µk . 5. The statistical input distribution must be known (at least up to the 6th central moment). 4.2.4 Proof: Nonlinear Functions Produce Non-Gaussian Output Distributions Consider the propagation of Gaussian error. With a Gaussian distribution, the fol- lowing expressions are true: • All odd moments (µk , where k is odd) are equal to zero • The fourth moment is equal to three times the second moment squared[39] (µ4 = 3µ22 ) • The sixth moment is equal to fifteen times the second moment cubed (µ6 = 15µ32 ) Consequently, Eq. (4.3) reduces to Eq. (4.4) when inputs are Gaussian. γ1 ≈ 3σx ∂12 ∂2 + σx3 ∂23 1.5 ∂12 + 12 ∂22 σx2 (4.4) where σx is the standard deviation of the input distribution. Equation (4.4) proves that nonlinear functions (i.e., the second partial derivative is non-zero) produces a skewed nonGaussian output, even with Gaussian inputs. Consequently, the commonly-made assumption that outputs are Gaussian is generally erroneous for nonlinear functions. 38 4.3 Propagation of Kurtosis Any statistical property that is propagated through a system model improves the accuracy of the predicted output distribution. For example, propagating both a mean and a variance is more accurate (and useful) than propagating a mean alone. In a similar manner, propagating skewness (as shown in Section4.2) in addition to a mean and variance also improves the accuracy of the predicted output distribution. Further improvements in accuracy can be obtained by also propagating kurtosis, the fourth-order statistic. This section defines kurtosis and excess kurtosis, and derives a formula for propagating kurtosis through an analytical system model. 4.3.1 Definition of Kurtosis The fourth-order statistic is kurtosis. Kurtosis is a measure of a distribution’s “peaked- ness,” or the thickness of the distribution’s tails. Kurtosis (denoted β2 ) is the fourth standardized moment, and is defined in Eq. (4.5). " β2 = E = x − x̄ σ µ4 σ4 4 # (4.5) The kurtosis of a Gaussian distribution is equal to 3. 4.3.2 Definition of Excess Kurtosis In statistical analysis, “excess kurtosis” (denoted γ2 ) is often used more than kurtosis. In practice, the term “kurtosis” more often refers to excess kurtosis instead of the fourth standardized moment. To avoid confusion, this thesis uses the definition of kurtosis presented above and defines excess kurtosis as the fourth cumulant divided by the square of the second cumulant, as indicated in Eq. (4.6). Cumulants (denoted κi ) are statistical properties similar to moments, where κ2 = µ2 , κ3 = µ3 , and κ4 = µ4 − 3µ22 . While it can be more convenient to 39 use cumulants instead of moments in some statistical analyses, the propagation of uncertainty using a Taylor series is much simpler and easier using central moments. Since a Gaussian distribution has a kurtosis of three, the “minus 3” in Eq. (4.6) causes a Gaussian distribution to have zero excess kurtosis. κ4 κ22 µ4 = −3 σ4 γ2 = (4.6) Figure 4.4 shows the excess kurtosis of several common types of distributions. Figure 4.4: Excess Kurtosis of Various Common Statistical Distributions 4.3.3 Kurtosis Propagation Formula Derivation A second-order Taylor series will also be used to propagate kurtosis through a system model. The third central moment of output y has already been given in Eq. (4.2), and the fourth moment is given in Eq. (4.7). 40 4 E (y − ȳ) ≈ µ4 ∂14 3 ∂1 ∂2 2 2 3 3 · ∂1 ∂2 (µ2 − 2µ2 µ4 + µ6 ) 2 3 1 3 2 ∂1 ∂2 (µ2 µ3 − µ2 µ5 + 3 µ7 ) 2 1 4 2 4 ∂2 (6µ2 µ4 − 3µ2 − 4µ2 µ6 + µ8 ) 16 2(µ5 − µ2 µ3 ) (4.7) The excess kurtosis γ2 in the output distribution y is given by Eq. (4.8). γ2 = β2 − 3 (4.8) where kurtosis β2 is given by Eq. (4.9). β2 E (y − ȳ)4 = 2 E (y − ȳ)2 4 µ4 ∂ 1 3 ∂1 ∂2 2(µ5 − µ2 µ3 ) 2 2 3 3 · ∂ ∂ − 2µ µ + µ ) (µ 2 4 6 2 2 1 2 3 ∂1 ∂23 (µ22 µ3 − µ2 µ5 + 31 µ7 ) 2 1 4 2 4 ∂2 (6µ2 µ4 − 3µ2 − 4µ2 µ6 + µ8 ) 16 ≈ 2 µ2 ∂12 + µ3 ∂1 ∂2 + 41 (µ4 − µ22 ) ∂22 (4.9) An estimate of the kurtosis of an output distribution can be obtained using Eq. (4.9) and a known input distribution. The input central moments µk can be estimated using any appropriate population sampling technique. 4.4 Example: Solar-Powered Unmanned Aerial Vehicle Propeller Thrust Consider the power and propulsion system of the solar-powered unmanned aerial vehicle (UAV) shown in Figure 4.5. 41 Figure 4.5: Power and Propulsion System of a Solar-Powered Unmanned Aerial Vehicle (UAV) Larson (see Reference [21]) and the author of this thesis have previously created a compositional system model for this system (as shown in Figure 4.6), deterministically propagated a max/min error envelop through the model, and validated the results using real data. As this thesis focuses on statistical, non-deterministic error analysis, a full description of this model and the deterministic approach used to propagate error will not be replicated in this thesis. However, the propeller component of this system and its corresponding data will be used to demonstrate the increased accuracy obtained from propagating higher-order statistics through a system model. 4.4.1 The Analytical Model of Thrust The thrust T that the propeller produces is given by Eq. (4.10)[48]. T = Ct ρω 2 D4 42 (4.10) Figure 4.6: Compositional System Model of a Solar-Powered Unmanned Aerial Vehicle (UAV) where Ct is a unitless coefficient of thrust calibrated to be 3.458e-3, ρ is the density of air assumed to be constant at 1kg/m3 , and D is the propeller diameter measured to be 0.1778m. The angular velocity of the propeller, ω, is provided in rad/s, and is determined by the motor speed, which in turn is determined by a motor throttle command. Since a given throttle command does not consistently produce an exact motor speed, there is some variation in the angular velocity of the propeller. This distribution in angular velocities is described in Table 4.2 and shown in Figure 4.7. Table 4.2: Statistical Properties of the Angular Velocity Distribution of a Propeller Statistical Property Mean (ω̄): Variance (σω2 ): Skewness (γ1 ): Excess Kurtosis (γ2 ): 43 Value 526.7 9.383 -1.289 4.896 Input Distribution 0.2 Prob. (%) 0.15 0.1 0.05 0 510 515 520 525 530 Angular Velocity ω, rad/s 535 540 Figure 4.7: Distribution of the Angular Velocities of a Propeller in a Solar-Powered UAV 4.4.2 Statistical Component Model Output Prediction Based on this distribution of ω, the first eight central moments were calculated using Eq. (3.2). These moments, given in Table 4.3, are used to propagate statistical properties from the angular velocity distribution to predict a distribution for thrust output by the propeller at a given motor throttle command. Table 4.3: Central Moments of the Distribution of Angular Velocity, ω Central Moment Value 1st Moment (µ1 ): 0 2nd Moment (µ2 ): 9.355 3rd Moment (µ3 ): -36.34 4th Moment (µ4 ): 6.617e2 5th Moment (µ5 ): -1.063e4 6th Moment (µ6 ): 2.726e5 7th Moment (µ7 ): -8.678e6 8th Moment (µ8 ): 3.308e8 44 The second-order prediction of the mean thrust output T̄ was calculated to be 0.9587N using Eq. (3.13). Equation (3.15) predicts a variance of 1.232e-4N2 . Typically, higherorder statistics are not propagated and a Gaussian output distribution is assumed, which is compared with actual thrust measurements in Figure 4.8. This prediction only accounts for 65% of the actual system output distribution (i.e., 65% overlap in the area under the predicted and actual probability density functions). Predicted Output Distribution from Mean/Var. Only Actual Output Distribution 0.8 0.5 Prob. (%) Prob. (%) 0.6 0.4 0.3 0.4 0.2 0.2 0.1 0 0.88 0.9 0.92 0.94 0.96 Propeller Thrust, N 0.98 0 1 (a) Gaussian Prediction Using Mean and Variance Only 0.9 0.92 0.94 0.96 Propeller Thrust, N 0.98 (b) Actual Thrust Measurements Figure 4.8: Predicted Gaussian Output Distribution of Thrust Obtained by Propagating a Mean and Variance Only (a) Compared with Actual Thrust Measurements (b) The methods presented in this chapter to propagate skewness and kurtosis, using Eqs. (4.3) and (4.9), respectively, results in a more accurate prediction of the system output. This predicted output, shown in Figure 4.9a, accounts for 79% of actual system outputs—a 22% improvement in accuracy compared with propagating a mean and variance alone. 4.5 Example: Flat Rolling Metalworking Process Consider the manufacture of steel plates or sheets via flat rolling, where material is fed between two rollers (called working rolls). The gap between the working rolls is less than the thickness of the incoming material. As the working rolls rotate in opposite directions, the incoming material elongates as its thickness is reduced. This process, illustrated in 45 Predicted Output Distribution from Mean, Var., Skew., & Kurt. Actual Output Distribution 0.8 0.5 Prob. (%) Prob. (%) 0.6 0.4 0.3 0.4 0.2 0.2 0.1 0 0.88 0.9 0.92 0.94 0.96 Propeller Thrust, N 0.98 0 1 (a) Non-Gaussian Prediction Using Mean, Variance, Skewness, and Kurtosis 0.9 0.92 0.94 0.96 Propeller Thrust, N 0.98 (b) Actual Thrust Measurements Figure 4.9: Predicted Non-Gaussian Output Distribtuion of Thrust Obtained by Propagating a Mean, Variance, Skewness, and Kurtosis (a) Compared with Actual Thrust Measurements (b) Figure 4.10, can be done either below the recrystallization temperature of the material (cold rolling) or above it (hot rolling). α R h0 v0 vf hf Figure 4.10: Flat Rolling Manufacturing Process Whereby Plates or Sheets of Metal are Made—Material is Drawn Between Two Rollers, which Reduces the Material’s Thickness 46 4.5.1 The Analytical Model The manufacturer desires to use its flat rolling equipment more efficiently by reducing overall rolling time for each plate. Consequently, the manufacturer desires to minimize the number of passes required to achieve final plate thickness. The maximum amount of deformation that can be achieved in a single pass is a function of the friction at the interface between the rolls and the material. If the intended change in thickness is too great, the rolls will merely slip along the material without drawing it in [49]. The maximum change in thickness attainable in a single pass (∆Hmax ) is given in Eq. (4.11) [50]. ∆Hmax = µ2f R (4.11) where µf is the coefficient of friction between the rolls and the plate, and R is the radius of the rolls. In this example, the radius of the rolls is measured to be 1m. Determining the coefficient of friction in a metalworking process is more difficult, however. The conditions surrounding friction in a metalworking process are very different from those in a mechanical device [49], as shown in Table 4.4. Table 4.4: Friction Conditions in Mechanical Devices and Metalworking Processes Typical Mechanical Devices -Two surfaces of similar material and strength -Elastic loads and no change in shape -Wear-in cycles produce surface compatibility -Low-to-moderate temperatures -Friction force depends on contact pressure Metalworking Processes -One very hard tool and one softer material -Plastic deformation occurs in material -Each set of rollers makes a single pass on material -Contact area constantly changes under deformation -Often elevated temperatures -Friction force depends on material strength Furthermore, lubrication is often used both to reduce friction and consequent tool wear, and to act as a thermal barrier to help regulate tool temperature [51]. All these 47 factors and others (e.g., rolling speed, material properties, surface finishes, etc.) combine to create variation in the friction experienced in the flat rolling metalworking process. This variation can inhibit the manufacturer’s ability to specify a gap width (and the resulting change in material thickness, ∆H) for each pass. While many empirical and mathematical formulas have been presented as methods to predict the coefficient of friction in flat rolling processes, these will not be addressed in this thesis. For the purposes of this example, it is sufficient to assume that some appropriate technique has been employed to determine the distribution of friction coefficients. This distribution is described in Table 4.5 and shown in Figure 4.11. Table 4.5: Statistical Properties of the Coefficient of Friction Distribution Statistical Property Mean (µ̄): Variance (σ 2 ): Skewness (γ1 ): Excess Kurtosis (γ2 ): Value 0.35 9e-4 0.7 0.2 Input Distribution Prob. (%) 0.1 0.05 0 0.25 0.3 0.35 0.4 0.45 Coefficient of Friction, µf 0.5 Figure 4.11: Distribution of the Coefficient of Friction in a Flat Rolling Metalworking Process 48 4.5.2 Statistical System Model Output Prediction Based on this distribution of µf , the first eight central moments were calculated using Eq. (3.2). These moments, given in Table 4.6, are used to propagate statistical properties from the friction coefficient distribution to predict a distribution for the maximum change in thickness attainable with a single pass. Table 4.6: Central Moments of the Distribution of the Coefficient of Friction, µf . Central Moment 1st Moment (µ1 ): 2nd Moment (µ2 ): 3rd Moment (µ3 ): 4th Moment (µ4 ): 5th Moment (µ5 ): 6th Moment (µ6 ): 7th Moment (µ7 ): 8th Moment (µ8 ): Value 0 8.99e-4 1.88e-5 2.58e-6 1.45e-7 1.45e-8 1.21e-9 1.2e-10 The second-order prediction of the mean maximum reduction in thickness ∆H¯max was calculated to be 12.5cm using Eq. (3.13). Equation (3.15) predicts a variance of 0.136cm2 . Typically, higher-order statistics are not propagated and a Gaussian output distribution is assumed. This Gaussian prediction is compared with actual system output in Figure 4.12. This prediction only accounts for 53% of the actual system output distribution (i.e., 53% overlap in the area under the predicted and actual probability density functions). The methods presented in this chapter to propagate skewness and kurtosis, using Eqs. (4.3) and (4.9), respectively, results in a more accurate prediction of the system output. This predicted output, shown in Figure 4.13a, accounts for 93% of actual system outputs—a large improvement over propagating a mean and variance alone. 49 Actual Output Distribution 0.2 0.15 0.15 Prob. (%) Prob. (%) Predicted Output Distribution from Mean/Var. Only 0.2 0.1 0.05 0.1 0.05 0 −0.05 0 0.05 0.1 0.15 0.2 0.25 Maximum Change in Thickness Per Pass, ∆ Hmax 0 −0.05 0 0.05 0.1 0.15 0.2 0.25 Maximum Change in Thickness Per Pass, ∆ Hmax (a) Gaussian Prediction Using Mean and Variance Only (b) Actual System Output Figure 4.12: Predicted Gaussian Output Distribution Obtained from Propagating a Mean and Variance Only (a) Compared with Actual System Output (b) Predicted Output Distribution from Mean, Var., Skew., & Kurt. 0.2 Actual Output Distribution 0.2 0.15 Prob. (%) Prob. (%) 0.15 0.1 0.05 0.1 0.05 0 −0.05 0 0.05 0.1 0.15 0.2 0.25 Maximum Change in Thickness Per Pass, ∆ Hmax 0 −0.05 0 0.05 0.1 0.15 0.2 0.25 Maximum Change in Thickness Per Pass, ∆ Hmax (a) Non-Gaussian Prediction Using Mean, Variance, Skewness, and Kurtosis (b) Actual System Output Figure 4.13: Predicted Non-Gaussian Output Distribution Obtained from Propagating a Mean, Variance, Skewness, and Kurtosis (a) Compared with Actual System Output (b) 4.5.3 Ramifications of Neglecting Higher-Order Statistics Using the traditional approach where only a mean and variance are propagated and a Gaussian distribution is assumed, the manufacturer would have only been able to reduce the material thickness by a maximum of 3cm per pass, in order to achieve a 99.5% success rate. However, using the methods presented in this chapter to also propagate higher-order statistics, the manufacturer can confidently reduce the material thickness by 7.9cm per pass. 50 This reduces the number of passes required to achieve the desired plate thickness by over two and a half times. As this example clearly indicates, the benefits of propagating higher-order statistics through a system model can be substantial. 4.6 Comments on Propagation of Higher-Order Statistics The variance in a system’s output can easily be predicted using a first-order Taylor series and knowledge of the input variance. However, having only a mean and variance and lacking any additional information about the output distribution, system designers often make the erroneous assumption that the output is Gaussian. This chapter has shown how inaccurate that assumption can be, even for very simple functions. By following the methods shown in this chapter, system designers can more fully describe an output distribution by also propagating higher-order statistics, such as skewness and kurtosis, though a system model. While sufficient for many physical systems, the approach to higher-order statistical error propagation presented in this chapter may not work for all types of system models, such as state-space models, Laplace transforms, and differential equations. Additional work is required to adapt the method presented for use with these types of models. 51 52 CHAPTER 5. CONCLUSION Error propagation and uncertainty analysis is a complex but important part of the system design process. Accurately predicting a statistical probability density function for system outputs can mean the difference between a successful system design and one that fails to meet design objectives. The methods provided in this thesis allow the system designer to accurately propagate fully-described statistical distributions through a system model without incurring significantly greater computational costs. The correction factors presented in this thesis for the efficient and accurate propagation of variance have only been determined for trigonometric, exponential, and logarithmic functions. Additional work is required to determine correction factors for other nonlinear, closed-form functions. The author has based this research on the assumption that a closed-form, analytical equation is necessary to use the methods presented in this thesis. However, it is the author’s belief that this same approach can be applied to any model (closed-form or not) where outputs can be obtained from given inputs and derivatives can be numerically evaluated. While non-closed-form models are outside the scope of this research, the author still believes the correction factors presented in Chapter 3 can be determined for non-closed-form equations if the truncation error is predictable, outputs are attainable, and derivatives can be determined. Additional work is required to determine what characteristics the model must have in order to have a predictable truncation error. If these methods could be used with models that are not closed-form, it would greatly increase the quantity and variety of models this research applies to. This would allow the designer to use these methods for more real-world engineering problems, including finiteelement models, dynamic models, transfer functions, state-space models, differential equations, software packages, “black-box” models, and others. 53 The formulas for propagating skewness and kurtosis that have been developed in this thesis have produced very accurate estimates of actual output skewness and kurtosis in the examples used in this thesis. However, it is possible that these estimates may not be perfectly accurate for other nonlinear system models. If this is true, then additional work is required to determine if correction factors can be determined that achieve higher-order accuracy in skewness and kurtosis propagation without increased computational cost. Lastly, future work could make the correction factors and skewness and kurtosis propagation formulas more accessible to design engineers. This could involve creating a table to be published online and/or included in textbook appendices. 54 REFERENCES [1] Murphy, B., 2009. “Early verification and validation using model-based design.” EDN, 54, July, pp. 39–41. 1 [2] Poolla, K., Khargonekar, P., Tikku, A., Krause, J., and Nagpal, K., 1994. “A timedomain approach to model validation.” IEEE Transactions on Automatic Control, 39(5), May, pp. 951–959. 1, 5 [3] Hamaker, H. C., 1995. “Relative merits of using maximum error versus 3(sigma) in describing the performance of laser-exposure reticle writing systems.” Proc. SPIE, 2440, May, p. 550. 1, 5 [4] Vardeman, S. B., 1994. Statistics for Engineering Problem Solving. ISBN: 0-534-92871-4. PWS Publishing Company, Boston, MA. 1, 4, 13, 14, 18 [5] Koch, P. N., 2002. “Probabilistic design: Optimizaing for six sigma quality.” In 43rd AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference, Paper Number AIAA-2002-1471. 3 [6] Glancy, C., 1999. “A second-order method for assembly tolerance analysis.” In Proceedings of the 1999 ASME Design Engineering Technical Conferences, no. DAC-8707 in DETC99, ASME/DETC. 3, 10 [7] Kiureghian, A. D., 1996. “Structural reliability methods for seismic safety assessment: A review.” Engineering Structures, 18(6), June, pp. 412–424. 4 [8] Thanedar, P. B., and Kodiyalam, S., 1992. “Structural optimization using probabilistic constraints.” Structural and Multidisciplinary Optimization, 4, pp. 236–240 10.1007/BF01742750. 4 [9] Melchers, R. E., 1999. Structural Reliability: Analysis and Prediction. Ellis Horwood Series in Civil Engineering. John Wiley & Sons, New York. 4 [10] Parkinson, A., Sorensen, C., and Pourhassan, N., 1993. “A general approach for robust optimal design.” Journal of Mechanical Design, 115(1), March. 4 [11] Chen, W., Wiecek, M. M., and Zhang, J., 1999. “Quality utility - a compromise programming approach to robust design.” Journal of Mechanical Design, 121, June. 4 [12] Chen, W., Sahai, A., Messac, A., and Sundaraj, G. J., 2000. “Exploring the effectiveness of physical programming in robust design.” Journal of Mechanical Design, 122(2), June, pp. 155–163. 4 55 [13] Su, J., and Renaud, J. E., 1997. “Automatic differentiation in robust optimization.” AIAA Journal, 35(6), pp. 1072–1079. 4 [14] Taguchi, G., 1992. Taguchi on Robust Technology Development: Bringing Quality Engineering Upstream. Asme Press Series on International Advances in Design Productivity. ASME Press, New York, August. 4 [15] Messac, A., and Ismail-Yahaya, A., 2002. “Multiobjective robust design using physical programming.” Structurla and Multidisciplinary Optimization, 23(5), pp. 357–371. 4 [16] Oberkampf, W. L., DeLand, S. M., Rutherford, B. M., Diegert, K. V., and Alvin, K. F., 2002. “Error and uncertainty in modeling and simulation.” Reliability Engineering & System Safety, 75, March, pp. 333–357. 4 [17] Halton, J. H., 1960. “On the efficiency of certain quasi-random sequences of points in evaluating multi-dimensional integrals.” Numerische Mathematik, 2, pp. 84–90 10.1007/BF01386213. 4 [18] Hammersley, J. M., 1960. “Monte carlo methods for solving multivariate problems.” Annals of the New York Academy of Sciences, 86, May, pp. 844–874. 4 [19] Hutcheson, R. S., and McAdams, D. A., 2010. “A hybrid sensitivity analysis for use in early design.” Journal of Mechanical Design, 132(11), November, p. 111007. 4, 17, 38 [20] Owen, A. B., 1998. “Latin supercube sampling for very high-dimensional simulations.” ACM Transactions on Modeling and Computer Simulation, 8(1), January. 4 [21] Larson, B., Anderson, T. V., and Mattson, C. A., 2010. “System behavioral model verification for concurrent design and modeling.” In 13th AIAA/ISSMO Multidisciplinary Analysis and Optimization Conference Proceedings, no. 9104 in AIAA-2010, AIAA/ISSMO. 5, 42 [22] Evans, J. W., Zawadzki, R. J., Jones, S. M., Olivier, S. S., and Werner, J. S., 2009. “Error budget analysis for an adaptive optics optical coherence tomography system.” Optics Express, 17(16), August, pp. 13768–13784. 5 [23] Hamaker, H. C., 1995. “Improved estimates of the range of errors on photomasks using measured values of skewness and kurtosis.” Proc. SPIE, 2621(198 (1995)), March. 5 [24] Oschmann, J., 1997. Gemini system error budget plan, January. 5 [25] Samet, H., 2005. Foundations of Multidimensional and Metric Data Structures (The Morgan Kaufmann Series in Computer Graphics and Geometric Modeling). Morgan Kaufmann Publishers Inc., San Francisco, CA, USA. 6 [26] Hayes, B., 2003. “A lucid interval.” American Scientist, 91(6), November-December, pp. 484–488. 6 [27] Box, G. E., and Tiao, G. C., 1992. Bayesian Inference in Statistical Analysis. ISBN: 9780471574286. Wiley, Hoboken, NJ, April. 6 56 [28] research.microsoft.com, 2011. http://goo.gl/rSJCD, July. 6 Basics of bayesian inference and belief networks [29] Box, G. E. P., and Wilson, K. B., 1951. “On the experimental attainment of optimum conditions.” Journal of the Royal Statistical Society. Series B (Methodological), 13(1), pp. 1–45. 6 [30] Chen, W., Baghdasaryan, L., Buranathiti, T., and Cao, J., 2004. “Model validation via uncertainty propagation and data transformations.” AIAA Journal, 42(7), p. 14061415. 6 [31] Lombardi, M., and Haftka, R. T., 1998. “Anti-optimization technique for structural design under load uncertainties.” Computer Methods in Applied Mechanics and Engineering, 157(1-2), pp. 19–31. 7 [32] Jackson, P. S., 1982. “A second-order moments method for uncertainty analysis.” IEEE Transactions on Reliability, R-31(4), Octobber, pp. 382–384. 10 [33] Shapiro, S. S., and Gross, A. J., 1981. Statistical Modeling Techniques. Marcel Dekker, Inc., New York and Basel. 10 [34] Walwyn, R., 2005. Moments - Encyclopedia of Statistics in Behavioral Science., Vol. 3 John Wiley and Sons, Ltd, 111 River St. Hoboken New Jersey United States. 11 [35] Taylor, B., 1715. Methodus Incrementorum Directa et Inversa [Direct and Reverse Methods of Incrementation]. Proposition VII, Theorem 3, Corollary 2. London Translated into English in D. J. Struik, A Source Book in Mathematics 1200-1800 (Cambridge, Massachusetts: Harvard University Press, 1969), pages 329-332. 11 [36] Mattson, C. A., and Messac, A., 2002. “A non-deterministic approach to concept selection using s-pareto frontiers.” Proceedings of ASME DETC, 2(DETC2002/DAC34125), September, pp. 859–870. 13, 16, 17, 38 [37] Tellinghuisen, J., 2001. “Statistical error propagation.” Journal of Physical Chemistry A, 105b(15), March, pp. 3917–3921. 13, 18 [38] Lindberg, V., 2000. Uncertainties and error propagation - part i of a manual on uncertainties, graphing, and the vernier caliper Internet, July. 13 [39] Julier, S., Uhlmann, J., and Durrant-Whyte, H. F., 2000. “A new method for the nonlinear transformation of means and covariances in filters and estimators.” IEEE Transactions on Automatic Control, 45(3), March, pp. 477–482. 16, 38 [40] Putko, M. M., III, A. C. T., Newman, P. A., and Green, L. L., 2002. “Approach for uncertainty propagation and robust design in cfd using sensitivity derivatives.” Journal of Fluids Engineering, 124(1), March, pp. 60–69. 16 [41] Hamel, J., Li, M., and Azarm, S., 2010. “Design improvement by sensitivity analysis under interval uncertainty using multi-objective optimization.” Journal of Mechanical Design, 132(8), August, p. 081010. 16 57 [42] Goodman, L. A., 1960. “On the exact variance of products.” Journal of the American Statistical Association, 55(292), December, pp. 708–713. 18 [43] Zhang, J., and Du, X., 2010. “A second-order reliability method with first-order efficiency.” Journal of Mechanical Design, 132(10), October, p. 101006. 18 [44] Aono, H., Shyy, W., and Liu, H., 2009. “Near wake vortex dynamics of a hovering hawkmoth.” Acta Mechanica Sinica, 25(1), pp. 23–36. 28 [45] George, R. B., 2011. “Design and analysis of a flapping wing mechanism for optimization.” Master’s thesis, Brigham Young University, Provo, Utah, June. 28 [46] Mekid, S., and Vaja, D., 2008. “Propagation of uncertainty: Expressions of second and third order uncertainty with third and fourth moments.” Measurement, 41(6), July, pp. 600–609. 34 [47] von Hippel, P. T., 2005. “Mean, median, and skew: Correcting a textbook rule.” Journal of Statistics Education, 13(2). 35 [48] Garner, W. B., 2009. Model airplane propellers rc.org/pdf/Model%20Propellers%20Article.pdf, March. 42 http://www.dc- [49] Degarmo, E. P., Black, J. T., and Kohser, R. A., 2003. Materials and Processes in Manufacturing., 9 ed. ISBN: 0-471-65653-4. Wiley. 47 [50] Tyfour, W. R. Rolling Internet (http://www.freewebs.com/tyfour/Rolling.doc). 47 [51] Lenard, J. G., 2007. Primer on Flat Rolling. ISBN: 0080453198. Elsevier Science. 47 58 APPENDIX A. STEP-BY-STEP SOLUTION TO DUAL-PROPELLER, THREE DEGREE-OF-FREEDOM HELICOPTER PROBLEM Consider the compositional system model for a dual-propeller, three degree-of-freedom helicopter discussed in Section 3.6. This model, given initally in Eq. (3.25), is repeated below in Eq. (A.1). θ̈ = C1 C2 + C3 (m2 L2 − m1 L1 )g C1 = m1 L21 + m2 L22 + Jy C2 = cos θ L1 km (uL + uR ) C3 = m1 L21 + m2 L22 + Jy (A.1) Recall that variation exists in variables m1 , m2 , L1 , L2 , Jy , and km , and θ. The means and standard deviations for these variables are repeated in Table A.1. Table A.1: Design Parameters for Dual-Propeller, 3-DOF Helicopter Design Parameter Mean m1 0.891 (kg) m2 1.000 (kg) L1 0.850 (m) L2 0.3048 (m) Jy 0.0014 (kg· m2 ) km 0.0546 θ 0 59 Std. Dev. 10−4 (kg) 10−4 (kg) 10−4 (m) 10−4 (m) 10−5 (kg· m2 ) 10−5 1◦ A.1 Variance Propagation Through Component 1 All first-order partial derivatives of component C1 , evaluated at the input means, are given in Eq. (A.2). ∂C1 ∂m1 ∂C1 ∂m2 ∂C1 ∂L1 ∂C1 ∂L2 ∂C1 ∂Jy −gL1 (m1 L21 + m2 L22 + Jy ) − gL21 (m2 L2 − m1 L1 ) = −5.4096 (m1 L21 + m2 L22 + Jy )2 gL2 (m1 L21 + m2 L22 + Jy ) − gL22 (m2 L2 − m1 L1 ) = 4.8085 (m1 L21 + m2 L22 + Jy )2 −gm1 (m1 L21 + m2 L22 + Jy ) − 2gm1 L1 (m2 L2 − m1 L1 ) = 0.5020 (m1 L21 + m2 L22 + Jy )2 gm2 (m1 L21 + m2 L22 + Jy ) − 2gm2 L2 (m2 L2 − m1 L1 ) = 18.2601 (m1 L21 + m2 L22 + Jy )2 −g(m2 L2 − m1 L1 ) = 8.1501 (m1 L21 + m2 L22 + Jy )2 = = = = = (A.2) The first-order Taylor series approximation from Eq. (3.9) is used to propagate variance through this component, as given in Eq. (A.3). 2 σC1 ≈ 2 n X ∂C1 i=1 ∂xi 2 σx2i ∂C1 2 ≈ σm1 + ∂m1 ≈ 3.8673e-6 ∂C1 ∂m2 2 2 σm2 + ∂C1 ∂L1 2 2 σL1 + ∂C1 ∂L2 2 2 σL2 + ∂C1 ∂Jy 2 2 σJy (A.3) The mean value for component C1 is -6.0152, which was determined by evaluated C1 at the input means. A.2 Variance Propagation Through Component 2 All first- and second-order partial derivatives of component C2 , evaluated at the input means, are given in Eq. (A.4). 60 ∂C2 = − sin(θ) = 0 ∂θ ∂ 2 C2 = − cos(θ) = −1 ∂θ2 (A.4) The second-order Taylor series approximation from Eq. (3.17) is used to propagate variance through this component, as given in Eq. (A.5). Note that the variance must be expressed in radians. 2 σC2 2 2 ∂C2 1 ∂ 2 C2 2 ≈ σθ + σθ4 ∂θ 2 ∂θ2 ≈ 4.6396e-8 (A.5) A correction factor e is then calculated using Eq. (3.19, as shown in Eq. (A.6). 1 2 1 + 1.022σC2 = 0.9997 e = (A.6) The resulting variance in component C2 is given by Eq. (A.7). 2 2 ≈ σC2 e σC2,CF ≈ 4.6381e-8 (A.7) The mean value for component C2 is 1, which was determined by evaluated C2 at the input means. 61 A.3 Variance Propagation Through Component 3 All first-order partial derivatives of component C3 , evaluated at the input means, are given in Eq. (A.8). ∂C3 ∂m1 ∂C3 ∂m2 ∂C3 ∂L1 ∂C3 ∂L2 ∂C3 ∂Jy ∂C3 ∂km = = = = = = −L31 km (uL + uR ) = −5.5401 (m1 L21 + m2 L22 + Jy )2 −L1 L22 km (uL + uR ) = −0.7124 (m1 L21 + m2 L22 + Jy )2 km (uL + uR )(m1 L21 + m2 L22 + Jy ) − 2m1 L21 km (UL + uR ) = −4.9566 (m1 L21 + m2 L22 + Jy )2 −2m2 L1 L2 km (uL + uR ) = −4.6744 (m1 L21 + m2 L22 + Jy )2 −L1 km (uL + uR ) = −7.6680 (m1 L21 + m2 L22 + Jy )2 L1 (uL + uR )(m1 L21 + m2 L22 + Jy ) = 103.6514 (A.8) (m1 L21 + m2 L22 + Jy )2 The first-order Taylor series approximation from Eq. (3.9) is used to propagate variance through this component, as given in Eq. (A.9). 2 σC3 ≈ 2 n X ∂C3 i=1 ∂xi 2 σx2i 2 2 2 ∂C3 ∂C3 ∂C3 ∂C3 2 2 2 2 ≈ σm1 + σm2 + σL1 + σL2 ∂m1 ∂m2 ∂L1 ∂L2 2 2 ∂C3 ∂C3 2 2 + + σJy σkm ∂Jy ∂km ≈ 1.8564e-6 (A.9) The mean value for component C3 is 5.6594, which was determined by evaluated C3 at the input means. 62 A.4 Variance Propagation Through System-Level Compositional Model All first-order partial derivatives of the system model, evaluated at the input means, are given in Eq. (A.10). ∂ θ̈ = C2 = 1 ∂C1 ∂ θ̈ = C1 = −6.0152 ∂C2 ∂ θ̈ = 1 ∂C3 (A.10) Now that the variance and mean of each component has been determined, the firstorder Taylor series approximation from Eq. (3.9) is used to propagate variance through the system-level compositional model, as given in Eq. (A.11). σθ̈2 ≈ ∂ θ̈ ∂C1 !2 2 σC1 + ∂ θ̈ ∂C2 ≈ 7.4019e-6 !2 2 σC2 + ∂ θ̈ ∂C3 !2 2 σC3 (A.11) Actual variance in θ̈, as determined from a Monte Carlo simulation with 10 million data points, is 6.5979e-6, meaning this compositional system model with a correction factor has 12% relatvie error. By comparison, the estimate of variance propagation using strictly a first-order Taylor series (without the compositional model or a correction factor) is 4.4853e-6 (32% relative error). As explained in Chapter 3, this increase in accuracy comes with very little additional computational cost. 63
© Copyright 2026 Paperzz