520 likes | 536 Views
This section introduces the concept of parameter estimation in statistical inference. It discusses the importance of sampling, selecting the appropriate distribution, and estimating the parameters. It also covers point estimation and interval estimation, along with the methods of moments, linear regression, and likelihood techniques.
E N D
Section 4: Parameter Estimation ESTIMATION THEORY - INTRODUCTION • The parameters associated with a distribution must be estimated on the basis of samples obtained from the population the distribution is modeling. The role of sampling as it relates to the statistical inference and parameter estimation is outlined in the figure in the next overhead. • The point is to construct a mathematical model that captures the population under study. This requires • inferring the type of distribution that best characterizes the population; and • estimating parameters once the distribution has been established. • Thus sampling a population will yield information in order to establish values of the parameters associated with the chosen distribution.
Tensile Strength Random Variable X Realizations of random variable X: 0 < x < Assume random variable is characterized by the distribution fX(x) ¥ + Familiar Statistical Estimators x = ( xi) n { (xi - x) } 2 s 2 = n - 1 Section 4: Parameter Estimation Experimental Observations - MOR bars or Tensile Specimens { x1, x2, ... , xn } Construct histogram to simulate fX(x) fX(x) x S Inferences on fX(x) S
Section 4: Parameter Estimation • “Choosing” a distribution can be somewhat of a qualitative and subjective process. We stress that the physics that underlie a problem should indicate an appropriate choice. However, most times the engineer is left with somehow establishing a rational choice, and too often histograms and their shapes are relied on. However there are quantitative tools that can aid the engineer in his/her selection. These tools are known as goodness-of-fit tests, e.g, • Anderson-Darling Goodness-of-Fit Test • Usually these types of tests will only indicate when the engineer chooses badly. For ceramics, the industry has focused on the two parameter Weibull distribution. This is a Type III minimum extreme value statistic. Thus physics and mathematics drive this selection. • Once the type of distribution has been made, the next step involves parameter estimation. There are two types of parameter estimation • Point Estimates • Interval Estimates
Section 4: Parameter Estimation Point estimation is concerned with the calculation of a single number, from a sample of observations, that “best” represents the parameters associated with a chosen distribution. Interval estimation goes further and establishes a statement on the confidence in the estimated quantity. The result is the determination of an interval indicating the range wherein the true population parameter is located. This range is associated with a level of confidence. For a given number of samples an increasing interval range increases the level of confidence. Alternatively, increasing the sample size will tend to decrease the interval range for a given level of confidence. Best possible combination is a large confidence and small interval size. The endpoints of the interval range define what is known as “confidence bounds.”
Section 4: Parameter Estimation POINT ESTIMATION - PRELIMINARIES • In general, the objective of parameter estimation is the derivation of functions, i.e., estimators, that are dependent on failure data, and that yield in some sense optimum estimates of the underlying population parameters. • Various performance criteria can be applied to ensure that optimized estimates are obtained consistently. Two important criteria are: • Estimate Bias • Estimate Invariance • Bias is a measure of the deviation of the estimated parameter value from the expected value of the population parameter. The values of point estimates computed from a number of samples will vary from sample to sample. In this context, the word sample implies a group of values, specifically a group of failure strengths. If enough samples are taken one can generate statistical distributions for the point estimates, as a function of sample size. If the mean of a distribution for a parameter estimate is equal to the expected value of the parameter, the associated estimator is said to be unbiased.
Section 4: Parameter Estimation • If an estimator yields biased results, the value of an individual estimate can easily be corrected if the estimator is invariant. An estimator is invariant if the bias associated with estimated parameter value is not functionally dependent on the true distribution parameters that characterize the underlying population. An example of an estimator that is not invariant is the linear regression estimators for the three-parameter Weibull distribution. The maximum likelihood estimator (MLE) for a two-parameter Weibull distribution is invariant. • There are three typical methods utilized in obtaining point estimates of distribution functions: • Method of moments • Linear regression techniques • Likelihood techniques
Section 4: Parameter Estimation METHOD OF MOMENTS
Section 4: Parameter Estimation • MINIMIZING RESIDUALS • No matter how refined our physical measurement techniques become, we can never ascertain the “true value” of anything. Thus we take repeated measurements of a quantity (say the distance between two corners of a property) and each time a measurement is conducted the values vary. Thus we are confronted with the dilemma of what value best represents the quantity measured. Several options include • Mean • Median • Mode • Faced with options, one should question which approach yields the “best possible” value. To answer this question a systematic approach is needed such that one can say • “This is the best possible value since this quantity is minimized” • or • “This is the best possible answer since that quantity is maximized”
Section 4: Parameter Estimation Thus we begin by focusing on the distance measuring example cited earlier and identify If many observations are made of this distance, then it is quite possible that none of the observations within a sample will coincide with the “best possible” value (whatever that is). If we define the difference between an observation and the “best possible” value as a residual where A residual is similar to, but not equal to an error. We need the true value to compute errors.
Section 4: Parameter Estimation We now search for a method to establish a “best possible” value. A systematic approach that yields the “best possible” value surely must minimize the residual associated with each observation (unless the observation is aberrant for some reason, i.e., the observation is an outlier). If we identify Then if the quantity S is minimized, the “best possible” value ( ) used in the computation above would have a quantifiable “goodness” associated with it, i.e., that the sum of the residuals has been minimized.
Section 4: Parameter Estimation To minimize the sum of the residuals, take the derivative of the expression above with respect to , set the derivative equal to zero and solve for Setting this last expression equal to zero definitely minimizes the residuals, for if no measurements are taken, all the residuals are zero. There is obviously a logic fault here.
Section 4: Parameter Estimation If minimizing the sum of the residuals is initially appealing (but the results do not help) then minimizing the sum of the squares of the residuals should be no less appealing. Here then Setting this last expression equal to zero yields Thus if we wish to minimize the sum of the squares of the residuals, then the sample mean should be utilized as the “best possible” value.
Section 4: Parameter Estimation Note that we developed this argument in terms of deriving a best possible value for a series of measurements. This concept can be easily extended to estimating values for distribution parameters, where instead of making a “measurement,” we take a sample from the underlying population. Minimizing the sum of the squares of the residuals is not the only systematic approach in producing the “best possible” estimates of distribution parameters. The maximum likelihood technique is another systematic approach where a “likelihood” is maximized. This technique is described in a later section. In some instances the estimators from various methods coincide, most times they do not. In situations where different approaches produce different estimators (and estimates), then one must choose between the different techniques. The amount of bias produced by an estimator is one measure of assessing efficacy. Additional statistical tools are available.
Section 4: Parameter Estimation PROBABILISTIC REGRESSION ANALYSIS We now wish to extend the concepts associated with regression analysis to parameter estimation. Consider an experiment where the tensile strength data has been collected for a given material. The tensile strength data is identified as the dependent variable (since the individual conducting the test can not control the value of this parameter – the material does). We need to adopt an independent random variable. Consider the ranked probability of failure associated with each tensile strength value depicted to the right. Note: Although the table above shows y dependent on x, in typical Weibull plots Pi is the vertical axis and strength is the horizontal axis. This causes confusion, especially using EXCEL.
Section 4: Parameter Estimation • Here • yi = ith ranked tensile strength • xi = Pi • = Associated ranked probability of failure • The data is ranked in the following fashion • y1 < y2 < y3 < ... < yn • Thus it seems reasonable to expect • P1 < P2 < P3 < ... < Pn • x1 < x2 < x3 < ... < xn • Note carefully that the individual conducting the experiment controls the value of n. This is important, as we will see later.
Section 4: Parameter Estimation • The ranked data is in ascending order. But what are the probability values associated with each ranked data value? Consider the following observations: • x1 corresponds to the lowest probability of failure • xn corresponds to the highest probability of failure • Assuming n is an odd integer
Section 4: Parameter Estimation To possibly account for these three observations, consider the following expression: For large n values,P1 trends to zero and Pn approaches 1. If we adopt this expression it is quite clear that the individual conducting the experiment influences Pi (or xi) through the choice of n prior to testing. Thus Pi (or xi) should be considered the independent variable in the experiment. With data collected from the experiment the individual analyzing the data now assumes an underlying probability density function If this expression can be linearized we can apply linear regression techniques to find the parameters q1 and q2.
Section 4: Parameter Estimation LINEAR REGRESSION – TWO PARAMETER WEIBULL DISTRIBUTION If we assume that the probability of failure in our experiment is governed by a two-parameter Weibull distribution, i.e. where s is the applied stress at failure, then this expression can be linearized as follows:
Section 4: Parameter Estimation If we take Then
Section 4: Parameter Estimation We can now make use of the traditional linear regression expressions for a and b Once a and b are determined the Weibull parameters and (where the tilde designates the value as an estimate) can be extracted from the expressions on the previous page.
Section 4: Parameter Estimation PROBABILITY OF FAILURE – RANKING SCHEMES A number of ranking schemes for Pi have been proposed in the literature. A mean ranking scheme was introduced in the previous section. In this section a median ranking scheme is discussed. As Johnson (1951) points out, the usual method of statistical inference involves constructing a histogram, from which a smoothed probability density function is derived. However, small sample sizes present difficulties since histograms vary greatly with changes in class intervals. As an alternative to this, ordered statistics were developed whereby ranked failure data is utilized. Consider a sample with five observations, where the observations are arranged in an increasing numerical order. It would seem reasonable to assume that the first observation (lowest value) would represent a value where 20% of the entire population would fall below this value. Thus an estimate for Pi of 20% is assigned this ranked value. Similarly a value of 40%, 60%, 80% and 100% would be assigned to the other ranked observations.
Section 4: Parameter Estimation If we concentrate on the first observation, assuming that 20% of the entire population falls below this value is a fairly far-reaching assumption. This statement can be made for the other four estimate Pi values Thus we will appeal to a statistical estimate of the population fraction that lies below this 20% value. To illustrate the concept, consider a sample of five observations taken from a population whose probability density function and attending distribution parameters are known. This sample of five is repeated four times, and for each sample the data is arranged in ascending order. If the cumulative distribution function for the first value is computed, then F(x1) = percentage of the population below the value of x1
Section 4: Parameter Estimation This is illustrated in the following figure taken from Lipson and Sheth (1979) Thus in sample #1 (darkened circle) the first failure may have occurred at A, where 15% of the population has a value less than value at A. For sample #2 (open circle), the first value occurs at B, which represents 9% of the population.
Section 4: Parameter Estimation When this procedure is repeated many times the data generates a series of percentage values that are randomly distributed. The median value of this distribution is given by the expression which equals when the total number of observations in the sample is n=5. From the second observation which equals for 5 observations.
Section 4: Parameter Estimation Thus in general Another ranking scheme proposed by Nelson (1982) had found wide acceptance. Here This estimator yields less bias then the median rank estimator, or the mean rank estimator. It is also the estimator accepted for use in ASTM 1239, and ISO Designation FDIS 20501. Due to client requests over the years WeibPar has included a host of other estimates for the probability of failure.
Section 4: Parameter Estimation METHOD OF MAXIMUM LIKELIHOOD The method of maximum likelihood is the most commonly used estimation technique for brittle material strength because the estimators derived by this approach maintain some very attractive features. Let (X1, X2, X3, …, Xn) be a random sample of size n drawn from an arbitrary probability density function with one distribution parameter, i.e., Here q is an unknown distribution parameter. The likelihood function of this random sample is defined as the joint density of the n random variables
Section 4: Parameter Estimation Often times it is much easier to manipulate the logarithm of the likelihood function, i.e., The maximum likelihood estimator (MLE) of q identified as , is the root of the expression obtained by equating the derivative of to zero If there is more than one parameter associated with a distribution, then derivatives of the log likelihood function are taken with respect to each unknown parameter, and each derivative is set equal to zero, i.e.,
Section 4: Parameter Estimation where And k represents the number of parameters associated with a particular distribution. When more than one parameter must be estimated often times the system of equations obtained by taking the derivative of the log likelihood function must be solved in an iterative fashion, e.g., as is done with the two parameter Weibull distribution inside WeibPar.
Section 4: Parameter Estimation The next two graphs illustrates how a first guess, and then a subsequent iteration affects the likelihood function. When then the first graph below plots this probability density function given above based on the first guess of m (=q1) and sq (=q2 ), which are admittedly pore choices.
Section 4: Parameter Estimation In both graphs the number of observations in the sample is n = 9. Note that all nine observed strength values fall to the right of the peak of the function. If the “sampling” procedure was truly random, the observed strength values would be more evenly spaced along the probability density function. This is why the first graph represents such a poor choice in the parameter estimates. Moreover, the likelihood function aids in quantifying whether or not the data is dispersed along the probability density function. To help visualize that the magnitude of the likelihood function does this, an arrow has been attached to the associated value of the probability function for each of the nine strength values. The value of the likelihood function would be the product of these nine values.
Section 4: Parameter Estimation Next consider the following graph which represents an iteration on the estimated distribution parameter values. Note the vertical scale has been maintained from the previous graph The shape and position of the probability density function appear to be a much better fit to the nine data points. Again, we base judgment on the assumption that our data values represent a random sample and they should therefore span the range. Note for small sample sizes this assumption can easily break down.
Section 4: Parameter Estimation • Again, nine arrows point to the associated values of the joint probability density function for each of the nine failure strengths. The product of these nine values represents the value of the likelihood function for this choice of distribution parameters. A simple inspection is sufficient to conclude that the likelihood from the latter iteration is greater than the likelihood from the former. If the latter choice of parameters is considered more acceptable, then this would indicate that obtaining a “best” set of distribution parameters involves maximizing the likelihood function. • Two important properties of maximum likelihood estimators • 1. Maximum likelihood estimators yield unique solutions • 2. Estimates asymptotically converge to the true parameters as the sample size increases
Section 4: Parameter Estimation MLE – TWO PARAMETER WEIBULL DISTRIBUTION Let s1, s2, ... , sNrepresent realizations of the ultimate tensile strength (a random variable) in a given sample, where it is assumed that the ultimate tensile strength is characterized by the two-parameter Weibull distribution. As noted in the previous section the likelihood function associated with this sample is the joint probability density evaluated at each of the N sample values. Thus this function is dependent on the two unknown Weibull distribution parameters (m, sq ). The likelihood function for an uncensored sample under these assumptions is given by the expression Here and designate estimates of the true distribution parameters m andsq .
Section 4: Parameter Estimation The system of equations obtained by differentiating the log likelihood function for a censored sample is given by And The top equation is solved first for . Subsequently is computed from the second expression. Obtaining a closed form solution for the first equation is not possible. This expression must be solved numerically. WeibPar solves these equations given a sample of failure data.
Section 4: Parameter Estimation EXAMPLE - SINGLE FLAW DISTRIBUTION WeibPar – Uncen.dat
Section 4: Parameter Estimation CONFIDENCE BOUNDS ON PARAMETER ESTIMATES
Section 4: Parameter Estimation MULTIPLE FLAW DISTRIBUTIONS (C.A. Johnson) The previous discussion assumes that single distribution of flaws is present throughout the specimens being tested. Real ceramic materials contain two or more types of defects, each with its own characteristic flaw size distribution. This can be easily verified using careful fractographic techniques. The other indication of multiple flaw distributions is a distinctive “knee” in the data. This is depicted in the figure to the right. This figure appears in ASTM C 1239. Here the probability of failure for the data points is estimated using
Section 4: Parameter Estimation The knee in the curve can appear as it does in the previous graph, or the curvature can be reversed as shown to the right. This figure appears in a GE report (1979) by C.A. Johnson. One can use the manner in which a knee is generated, i.e., how the data deviates from linearity to deduce information concerning the nature of the flaw distributions present. Consider a group of test specimens that when failed, have two distinctly different types of fracture origins. Some specimens fail from flaw type “A” while the remainder fail from flaw type “B.”
Section 4: Parameter Estimation • There are at least three ways in which these two flaw populations are present in the test specimens, i.e., the sample. • Both flaw distributions are present in every test specimen. This is known as “Concurrent Flaw Populations.” • Within a group of specimens any given specimen may contain flaws from distribution A, or from distribution B, but not from both. This is known as “Mutually Exclusive Flaw Populations.” • Within a group of specimens, flaw distribution A may be present in all specimens, while distribution B may be present in only some test specimens. This is known as “Partially Concurrent Flaw Populations.” • A larger number of cases must be considered than three above if more than two flaw distributions are active.
Section 4: Parameter Estimation • Examples of the three types of flaw distributions are as follows. • Concurrent Flaw Distributions – A group of test specimens machined from a ceramic billet which contains defects distributed throughout the volume. Each specimen then contains both machining flaws on the surface and volume defects interior to the test specimen. • Exclusive Flaw Distributions – A group of test specimens comprised of two subgroups purchased from two different manufacturer. • Partially Concurrent Distributions - A group of pressed and sintered test specimens which were all sintered in the same furnace, but could not be sintered in a single furnace cycle. The specimens sintered at the longer cycle would contain elongated grains (considered a deleterious flaws) and all the specimens would have strength degrading inclusions present from the powder preparation. • The parameter estimation methods presented here assumes that if multiple flaw populations are present, then they are concurrent flaw distributions. As indicated above, partially concurrent distributions point to processing methods that have not , or are not mature. In addition, it is quite typical for a manufacturer of ceramic components to identify one material supplier, and this would minimize the presence of exclusive flaw distributions.
Section 4: Parameter Estimation MLEs FOR MULTIPLE FLAW DISTRIBUTIONS The likelihood function for the two-parameter Weibull distribution where concurrent flaw distributions are present (known as a censored sample) is defined by the expression This expression is applied to a sample where two or more active concurrent flaw distributions have been identified from fractographic inspection. For the purpose of the discussion here, the different distributions will be identified as flaw types A, B, C, etc. When the expression above is used to estimate the parameters associated with the A flaw distribution, then r is the number of specimens where type-A flaws were found at the fracture origin, and i is the associated index in the first summation. The second summation is carried out for all other specimens not failing from type-A flaws (i.e., type-B flaws, type-C flaws, etc.). Therefore the sum is carried out from (j = r + 1) to N (the total number of specimens) where j is the index in the second summation. Accordingly, σi and σjare the maximum stress in the ith and jth test specimen at failure.
Section 4: Parameter Estimation • The system of equations obtained by differentiating the log likelihood function for a censored sample is given by • and • where: • r = the number of failed specimens from a particular group of a censored sample
Section 4: Parameter Estimation EXAMPLE - MULTIPLE FLAW DISTRIBUTIONS
Section 4: Parameter Estimation NON-LINEAR REGRESSION ANALYSIS: THREE PARAMETER WEIBULL DISTRIBUTION When strength data indicates the existence of a threshold stress, a three‑parameter Weibull distribution can be employed in the stochastic failure analysis of structural components. However, care must be taken such that careful fractography does not indicate the presence of a partially concurrent flaw distribution, which has the appearance of a population characterized by a three parameter Weibull distribution. The resulting expression for the probability of failure of a component fabricated from a material characterized by a three parameter Weibull distribution and subjected to a single applied stress is if the defect population is spatially distributed throughout the volume. A similar expression exists for failures due to area defects.
Section 4: Parameter Estimation Here the distribution g has the physical interpretation of a threshold stress. Regression analysis postulates a relationship between two variables. In an experiment typically one variable can be controlled (the independent variable) while the response variable (or dependent variable) is not. In simple failure experiments the material dictates the strength at failure, indicating that the failure stress is the response variable. The ranked probability of failure (Pi) can be controlled by the experimentalist, since it is functionally dependent on the sample size (N). After arranging the observed failure stresses (s1, s2, s3, , sN) in ascending order, and as before specifying then clearly the ranked probability of failure for a given stress level can be influenced by increasing or decreasing the sample size. The procedure outlined here adopts this philosophy. The assumption is that the specimen failure stress is the dependent variable, and the associated ranked probability of failure becomes the independent variable.
Section 4: Parameter Estimation Using the three-parameter version for the probability of failure, an expression can be obtained relating the ranked probability of failure (Pi) to an estimate of the failure strength ( ). Assuming uniaxial stress conditions in a test specimen with a unit volume, then where , and are estimates of the shape parameter (a), the scale parameter (b0), threshold parameter (g), respectively. Defining the residual as where siis the ith ranked failure stress obtained from actual test data, then the sum of the squared residuals is expressed as
Section 4: Parameter Estimation Here the following notation has been adopted where Note that the forms of and Wi change with specimen geometry. It should be apparent that the objective of this method is to obtain parameter estimates that minimize the sum of the squared residuals. Setting the partial derivatives of the sum of the squared residuals with respect to , and equal to zero yields the following three expressions
Section 4: Parameter Estimation and in terms of the parameter estimates. The solution of this system of equations is iterative, where the third expression is used to check convergence at iteration.
Section 4: Parameter Estimation EXAMPLE – THREE PARAMETER WEIBULL DISTRIBUTION