Mod-01 Lec-39 SEM - Measurement Model

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
Today, we will start a structural equation modelling part 1; that is measurement model. SEM, that is structural equation modelling, measurement modelling, measurement model. Now, let me see the content of today’s presentation. That we will start with conceptual model, then the assumptions of the model, then how to estimate the model parameters and model adequacy test, followed by a case study. You see in last class, I have explained that measure structural equation modelling has two components, one is measurement component, and another one is structural component. Measurement component is essentially a confirmative factor analysis, and structural part or we can say the structural model is equivalent to your path model or path analysis. Both the model those measurement as well as structural path, there are three important steps, one is model identification, then parameter estimation and model adequacy test. This is true for structural part, also in this lecture we will consider the measurement part which is confirmatory factor analysis. So, let us start with a conceptual model first. You see here. In last class I have shown you similar diagram and you see there are three factors Xi 1, Xi 2 and Xi 3 and each of the factors are manifested by a different variables starting from X 1, X 2, X 3 for Xi 1, X 4 and X 5 for Xi 2 and X 6, X 7, X 8 for Xi 3. So, in confirmatory factor analysis the basis is that that there are hidden constructs which are this or latern construct. Other you can say hidden variable that Xi 1, Xi 2, Xi 3 which are the causes of some manifest variable like X 1, X 2, X 3 to X 8 by putting this arrogate. For example, from Xi 1 to X 1, Xi 1 to X 2 and Xi 1 to x 3 we are restricting the model here in such a sense that we know that Xi 1 is manifested by X 1, X 2 and X 3. This manifest variable. Similarly, Xi 2 is manifested by X 4 and X 5. Similarly, Xi 3 is manifested by X 6, X 7 and X 8, another issue here that this hidden construct or latren construct, latern variable they co vary in the sense that if Xi 1 change there may be change of Xi 2. There may be change of Xi 3 where the correlation components is there. So, these type of, this is a typical structure of confirmatory factor model and other issues here, apart from this correlation between the construct which is denoted by phi. This come phi 2 1, this is phi 3 1, this curvature line, this curvature line is basically phi 3 2. Now, we are saying X 1 is caused by Xi 1 and as a result a causal linkage is given, Xi 1 to X 1 and parameter which basically depicting the relationship between Xi 1 and X 1 is lambda 1 1. Earlier, I also told you this lambda 1 1, this 1 1 this suffix comes that X 1 from X 1 this point is taken, Xi 1 1 is taken. So, it is not possible that the variability of X 1 will be fully explained by Xi 1. So, here is possibility of some other variables or hidden causes which may affect X 1 or we can say the errors part, noise part, all those things are considered by delta 1. So, in the same manner you have to explain X 1, X 2, X 3 and up to X 8. What I said verbally, this is depicted in equation form. We are saying X 1 is represented by lambda 1 or Xi 1 plus delta 1. If you think of regression line point of view you will be getting this equation. So, in the same manner there are as there are 8 X variables, so you are getting 8 linear equation and collective if you write in matrix form then that will be X 8 cross 1 equal to that capital lambda, which is a matrix of the dimension 8 cross 3. And which you can see here 8 cross 3 dimensions here, because Xi 1, Xi 2, Xi 3 that 3 and 8 X variables, this 8 plus this is Xi plus delta this one. So, this is the, this is the matrix form equation for this particular example. Now, you can go for a general equation from there that means what I means to say here the general equation will be X p cross 1 where we are saying. There are P number of manifest variables and which can be represented by this manner, the lambda which is a matrix of matrix relating P manifest variable to m factors. And this m factors are denoted by like this m cross 1 and definitely for every manifest variable there will be an error. So, this is the equation for confirmatory factor model, if you see recall that factor analysis you have found out there also similar relationship, but there are there is difference in the structure of the model then the model assumption in the covariance structure. So, what are the assumption? Here we assume that the expected value of delta equal to 0, that mean the noise variable, the error terms that mean is 0 and covariance of delta, this one is your theta. Or you can write theta delta, also m times theta delta and it is symmetric. So, if I say like this will be P cross P. So, this you variance component of delta of diagonal in the covariance that will be equal, that is why symmetric and delta is multivariate normal with mean 0 and covariance matrix theta delta. Or you can write theta also another important issue here is that assumption is that covariance between Xi, that mean manifest, sorry latern construct and the error term related to the manifest variable, they are 0. So, this is our your assumptions related to confirmatory factor model. Now, there as I told you there are covariance. So, the covariance structure of your see covariance structure, first one is covariance of X, this will be capital sigma P cross P. There will be covariance structure for the Xi latern construct, this one will be phi which is again m cross m matrix diagonal, not diagonal this is symmetric matrix. Then covariance of delta which we say theta P. P it is mostly assumed as diagonal matrix, assumed as diagonal not necessary always it will be diagonal,but it is assumed like this So, you have seen earlier in your exploiting factoring, also we have written X equal to that delta Xi plus, I think we had lambda Xi plus delta, means this capital lambda, replace delta. Now, if you find out the covariance of X then ultimately what you will be finding out? You will be finding out something like this plus covariance of this delta. This is theta in X and this phi was not there in excusive factor, it was I if phi equal to I then it is orthogonal factor analysis. So, this is null set that covariance structure and the relationship between this, okay? So, now let us stick to come to the model identification part, what I mean to say here by model identification if you clearly look into the model and the parameters that to be estimated as well as the information, what is available. There must be sufficient necessity and sufficiency of the information available to estimate the parameters of the confirmatory factor model. So, in order to do so you should now let us find out that what are the parameters, we require to be estimated if you see this slide you will see that we have few parameters to be estimated from in C F A. One is your lambda, capital lambda which is p cross m matrix. So, these many parameters to be estimated, there is phi which is also a p cross p matrix which is p cross p matrix, but being symmetric it has phi is m cross m matrix. I am sorry phi is m cross m matrix which is symmetric matrix. So, numbers of parameters will be m into m plus 1 by 2 to be estimated and then there is theta delta, as I told you that theta delta. What is this theta delta? This is delta, this is theta delta. So, theta delta or theta so there if we assume that it is diagonal then there will be p number of parameters to be estimated. So. we require to estimate capital lambda, we require to estimate phi, we require to estimate theta delta. Here in case of capital lambda p cross m, this number of parameters phi m into m plus 1 by 2, this number of parameter theta can number or parameters are there. Then in total the number of parameters to be estimated, number of parameter to be estimated we can write p m plus m into m plus 1 by 2 plus p, okay? Now, what you require to know that if t, suppose t equal to 50. If we require to estimate p 50 parameters and you require at least 50 simultaneous equations, getting me? Now, what information we have in case of confirmatory factor analysis? We have only one information, that is capital sigma which is the variance, co variance matrix of X, that is co variance matrix of X. Now, how many unique or non-redundant elements, this is also symmetric matrix. So, it has number of know redundant element equals to p into p plus 1 by 2. Now, see this what type of situation will occur? We require to estimate p number of parameters, it may so happen that number of parameters to be estimated is greater than number of independent non redundant elements in p. That is the available information, it may so happen that t equal to p into p plus 1 by 2. It may so happen that t less than p into p plus 1 by 2. Now, the first case, this is model cannot be identified. This is model is unidentified or under identified, un identified this case number of parameters or the number of unknows and knows are equal. This is uniquely identified case and this case this is over identified, because we have more information available over identified case. So, at least this two are necessary if you have this case type of situation which is uniquely estimated. If you have this over estimation, that is the desirable one over estimate is the desirable one. This condition, particularly this two, if this two conditions are either of the two is satisfying you are saying that necessary condition is satisfied. Necessary condition which is also known as ordered condition, okay? But ordered condition alone is not sufficient, this is necessary condition necessity is satisfied. There is another condition called rank condition, because then you will see that we have basically talking about the matrices. So, the rank condition, rank of my matrix is important issue here and it is little bit complicated one also. Rank conditions also verbally you have to satisfy then the rank condition is the sufficient condition and for example, is given that Willey in 1973 that is the reference. Now, order conditions satisfy the necessity and rank conditions satisfied the sufficiency. Let us assume that it is done. In the sense model e is identified, if model is identified. The next step is how to estimate the model. So, estimation of model parameters, okay? So, you have seen that we assume that ultimately that X, the manifest variable is normally distributed. It is the primitive multivariate normal p X variable p, p number of variable are there, multivariate normal with mean 0 and variance, co variance matrix sigma, capital sigma. Then for any observation multivariate observation Xi you can write that this is the PDF can be written like this, 2 pi to the power p by 2 sigma determinant to the power half e to the power minus half. Then X minus mu e, e is to write that is X i minus mu e is zero here. So, X minus m transpose, that means Xi transpose sigma inverse Xi, this is the multivariate normal distribution per density function for a particular multivariate observation. Now, we collect in observation i equal to 1 to n, we want to know the log first the likelihood. So, likelihood if you see this equation you find, you see the there is only one parameter which is sigma and mu e is 0. So, only one parameter is there, so we can write log of likelihood of sigma, not log likelihood of sigma which can be written as multiple equation of this, i equal to 1 to n f xi which will be multiplying i equal to X equal to 1, X 1, X 2 up to X n. Then the resultant will be like this 1 by 2 pi to the power n p by 2. Then determinant to the power n by 2, then e to the power minus half, then sum i equal to 1 to n. Then Xi transpose sigma inverse Xi that is what will be the likelihood and it is customary to take log likelihood. So, if you take log likelihood then what you get? You get minus m p by 2 log 2 pi for this term, minus n by 2 log this for this term minus half, i equal sum of i equal to 1 to n Xi transpose in sigma inverse Xi. So, obviously this from our we want to estimate this sigma that parameters it will be will go for some optimization root and there this constant term in the equation has whether you had keep it or do not keep it this is immaterial. So, we remove this constant. So, you can write this as minus n by 2 log of this plus half of I can write like this, minus n by 2 into this minus. So, minus half of this Xi transpose this xi. So, let me write a phrase that log of l, this sigma equal to minus n by 2 log determinant sigma minus half Xi transpose sigma inverse Xi. Now, this term can be written like this minus n by 2 log this minus, if I write n by 2 then summation i equal to 1 to n. This can be written like this, this can be written like this 1 by n, I am teaching because I have considered n here. So, this one can be written like this, write again, you write like this like this, then I come to R. R form n by 2, this minus n by 2 we can write this quantity as trace 1 by trace of 1 by xi transpose xi sigma inverse this, this is possible. Now, 1 by n Xi transpose Xi, this is nothing but variance, covariance matrix of the sample provided, n is large, 1 by n minus 1 and 1 by n become same. So, with modelling equation we can write like this minus n by 2, then this is trace of s sigma inverse. This one is minus n by 2 log of this plus trace of s sigma inverse. So, this is your log likelihood. Now, see what is the condition here in our estimation here, actually this procedure is like this, you will from the model from the model covariance of X, your lambda phi, lambda transpose plus theta delta. From this you will get covariance X sigma, that is sigma in terms of model parameter you collect sample, then you get the covariance matrix. Also, from sample there will be S. S is the again p cross q that sample covariance matrix, there what we want to do? We want to match this two, this two suppose a condition is such that S is equal to sigma, then if I put here what I can write? Log of l of s, this can be written like this, this log of S determinant of S plus trace of S S inverse. Now, S S inverse is I, so you can write like this. This log of S plus sum of the diagonal elements of the matrix I, that is p this is your equation number 1. Now, another one is we have already seen. The likelihood one, this equal to minus n by 2 log of this plus trace of this, this is our equation 2. So, what you want? We want to find out parameters. Now, this sigma this will be in terms of model parameter S here, and here when we are talking about S it is basically the numerical values and here it is in terms of model parameters like lambda phi and all those things. So, we will create a function now that we want to minimize that we are saying F theta which is nothing but S minus sigma theta of this nature which I am saying, not exactly which will be of this nature. So, then we can write like this F theta equal to log of l S minus log of l sigma, which if you write this is minus n by 2 log of determinant of S plus p plus n by 2 log of determent of sigma plus trace of S, this trace of S sigma inverse. So, this one you can write minus that n by 2. Then log of determinant of sigma plus trace of S sigma inverse minus log of S determent S minus p, we want to minimize this function. So, keeping this constant n by 2 again is of no use. So, final equation will be for our estimation, is this log of determinant of sigma plus trace of S sigma inverse minus log of S minus p. This is the equation, which want to minimize, okay? So, its null issue you have to use Newton Rapson or similar method, Newton Rapson similar method of numerical that optimization part. So, this is what is in the nut shell, the parameter estimation in place of confirmatory factor analysis, which is basically our measurement model, you see here. This is the theta log sigma theta trace of this, this is the case ignoring the constant. So, ultimately minimize this one using Newton Rapson or Gauss algorithms, okay? Now, let us see that whatever mathematics we have described now can it be put into a case study as a real life example. I will show you the example here which I have shown you earlier also in the in last, in the first class of structural equation modelling. I have shown you this one, but there what I have done actually, I have given you a glimpse of this things, just like scrawling down the slide. Because just the glimpse of what is this not will describe in detail that what is this measurement model and with the same case study, okay? So, the case study as you know it is a role of personal and socio technical factors in work injuries in mines and a study based on employee’s perceptions and you can see that the source is Paul and Maiti, 2008. It is published in the ergonomics, okay? So, let us start like this we have several manifest variables here. There are 18 manifest variables for example, age, experience, impulsivity, negative, affectivity, depression, risk taking, safety training, safety practice, safety equipment, availability, maintenance, job stress. Like this we are wondering that how I am saying these are manifest variable although, most of the things cannot be observed. So, actually what happen for every of the variable, we ask several questions and then those questions are summed into a particular quantity. And then that summed up values we have taken as the value for each of the variable, for each of the observations or individuals who participated in this study. So, in that sense it is manifested, means observed in that sense. Otherwise, its two layer questions actual it was like this, only one questions, then there sum to this. This manifest variables what we are saying there sum then further level of aggression actually. Suppose, there are question 1 to let it be 150 then there are manifest variable like Xi 1, like there are 18. Then this again it is aggregated into what I can say these are X 1, sorry these are X 1 to X 18 related to Xi 1 to some Xi, let it be Xi 9. So, this level of aggression is done here. So, we are taking in this level of aggregation, we are considering here this is manifest variable, but you may start from here to here. That will be combos some we have this. So, the same thing if you write in the confirmatory factor analysis form, it will be something like this. You see all the this covariance structure between this 9 gita Xi variables, it is not pictorially shown because of space concept. Otherwise, this will be this and then what will happen? We will immediately, you can go for the equations also for this, getting me? So, like I am giving one equation only here if I want to know what is X 1? Then this is nothing. I can write lambda 1 1 Xi 1 plus delta 1 if you consider this. So, here lambda 1 1 Xi 1 plus delta 1. Similarly, this one lambda 2 1 Xi 1 plus, so x 2 will be lambda 2 1 Xi 1 plus delta 2, if you consider X 3. So, X 3 is ultimately it is the single indicator manifest variable for the constant Xi 2. So, X 3 can be written like this that 3 2 xi 2 plus delta 3. So, in the same manner as there are X 18 so you will be able to write X 18, come to this one, X 18 is again a single indicator for Xi 9. So, this is your lambda 18 9 xi 9 plus delta 18. So, you can write in matrix form, when you write in matrix form you will be getting a equation in matrix form, equation we said X equal to lambda Xi plus delta. This equation you can find out here, here we have 18 cross 1. This will also be 18 cross 1, this is our 9 cross 1. So, this will be 18 cross 9, so this type of equation you can find out, okay? Now, let us see the model identification for this case. You just see that lambda part that how many lambdas are there. You count 1, 2,3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18. So, 18 lambdas, so we have written lambda 18 because others are 0. For example, lambda 3 1 if you give one linkage here lambda 3 1 that is 0, because it is a confirmatory. We know what are the manifest variable coming out of the hidden constructs phi. There are how many X 18. So, how many Xi 9 xi. So, m into m cross m into m plus 1 by 2 m into m plus 1 by 2. So, 9 into 10 by 2, that will be 45. So, you have lambda that is your 18, then your phi related variables will be 45 phi, related parameters will be 45 theta delta, again 18 delta 1 to delta 18. So, our t is 18 plus 45 plus 18 that is 81. Now, what is the unique elements? There we have 18 manifest variable cross 18. So, 18 into 19 by 2 this will be 171. Now, t equal to 81 it is much less than 171, the model is over identified, it is a good case so and necessity condition you satisfied and sufficiency we have not tested here that says the software they test all those things. Now, let us see that data part. The data part is actually random independent sampling. First we have taken accident group of workers followed by with frequency matching, non accident group of workers where all together 300 observations. So, immediately as I told you that how many independent non redundant element in your sigma matrix. This is the case, this is from sample, these are all co relation matrix. Now, question is what we want? We want basically to minimize this function and in this matrix is basically for S, that is sample co variance matrix, but actually we have taken correlation. Now, question comes whether co variance or correlation it all depends on the purpose of the study. In our purpose of the study we are more interested in the pattern of the relationship, then the original strength of relationship between latent variable and your manifest variable. We are more interested in the pattern of the relationships and not the original value. So, when you are interested in the pattern of the relationship R is a variable, R matrix should be used, that is correlation matrix should be used. Now, through definitely this is F theta, this is the function which is to be minimized and we have used this software, in this case LISREL linear structural relations. So, this software we use and ultimately this is what is the all the parameters which is estimated. You can see if you go back lambda 1 1, lambda 2 1, lambda 3 2, lambda 4 3 like this and you see that lambda 3 2. What is the value of lambda 3 2? Here, lambda 3 2 is 1 because this one indicator with one constant, we assume that this is the manifest variable itself is the construct, understood? So, that will be the what is the output of this measurement model. See ultimately we are talking about long back I think this one, these things when we have clubbed into this factors, these are all the factors or constructs latent constructs. This Xi 1 to Xi 9 these are not arbitrary Xi 1 to Xi 9, they have some meaning. Actually X 1, X 2 if you see that age and experience then demographics, this is the this is Xi 1 impulsivity negativity all four are clubbed to the var and Xi 2 value. Actually the negative personality is given here. I think this Xi 3, X 1, X 2, X 3. X 3 is work injuries. It is kept as it is what X 3, Xi 3 is the negative personality, then Xi 4 is your safety environment. Again Xi 5 is job stress, Xi 6 is social support, Xi 7 is job dissatisfaction, Xi 8 is work hazards, Xi 9 is safe work behaviour. So, these are all latent variables in this sense now we also want to know the co relation matrix between latent variables which is the output of this measurement model. You see that demographic work injuries negative personality, these are the latent constants and this is your correlation matrix in then there are little star is there. This star indicates point 0.05 probability level of significance. I think all are significant here except this value 0.04, 0.01 some other value, but essentially we are interested from measurement model to know that what is the correlation matrix of the latent constructs or factors what we are going to evaluate or estimate, getting me? That is what we have done and we have done this with the help of this, this lambda values and the Xi and the error term you are getting this values. And this will be this is a value, you want this is very, very important one because this will be used in structural model as input to structural model in structural equation modelling, I told you. That structural equation modelling two parts SEM has two parts, measurement model and structural model. The output of this will be input to this, fine? That I would consider. The structural measurement model would I consider measurement model as I said or not. If the model is not fit, if it is not adequate enough then the correlation matrix between the constructs generated they are not good. Also, we have doubt about those correlation values, we cannot abruptly accept this one. Now, in model adequacy test, in last class also I told you that fit index there are three types of fit index, absolute fit index, relative fit index and parsimonious fit index. Under absolute fit index chi square, chi square degree of freedom. So, absolute fit index, relative and parsimonious we are discussing about and absolute fit index there are many indices like chi square, chi square by degree of freedom, goodness of fit index, root mean square, residual root mean square. That is RMR RMSEA standard error approximation,then relative fit index. These are the standard indices available in any literature related to structural equation modelling and most of the, why most, I think almost all the indices are based on chi square value. So, we will discuss little of this. For example, absolute fit index what it does? It answers this question is the residual or unexplained variance remaining after model fitting appreciable. So, we do something like this. There will be two that hypothesis null and alternate hypothesis. Null hypothesis is we are saying that sigma equal to sigma theta, that actually that what you have estimated, that is correct and alternatively you are saying no, they are not correct. So, then this we will define one quantity called chi square, which is n minus 1 into F theta, that F theta you have seen that the minimization function, so that value you have after estimation. So, chi square that n minus 1 F theta, this follows this chi square distribution with nue degrees of freedom, where nue can be estimated like this, p into p plus 1 by 2 minus t, that is number of non-redundant elements minus number of parameters to be estimated. That is what more degrees of freedom available here and you will find out that what chi square value you get that should be as small as possible, because if for perfect fit F theta will be 0, the n minus 1 into F theta that it will be 0. So, 0 is the ideal value, but it will all depends on sample size, also n minus 1. Now, see that you will never get this your theta will be 0, because you are doing the numerical way of optimization, numerical optimization where some convergence value will be there. Now, if n is sufficiently large what will happen? This value will become large. So, what you want to how do then justify that whether the model is fit or not. One way is that this value should be as small as possible and other one is you go by chi square by degree of freedom. So, it is recommended in the later lecture that essentially the chi square distribution is such that the expected value of chi square nu is nu, because that is the degree of freedom. Because it is a this parameter in chi square we use the degree of freedom only. So, actually the chi square by the degree of freedom should be 1, but is not recommended what is said that 225 is the recommended value above 35 constraints. Now, if you use G F I that is goodness of fit index which is similar to R square in multiple regression, you can remember or recollect that R square equal to 1 minus SSE by SST. Now, you see this formulation here, that way we have written here that 1 minus trace of this by this. So, this is total variability and this one is the error term. It is similar to R square and this G F I value varies from 0 to 1 and it is desirable that G F I is greater than 0.90. So, in your model when develop a measurement model the software will give you the G F I value, if you find out that the G F I value is 0.9 or more, that it is good. It is desirable, but if it is less than 0.9 what you will do? You will not consider the error model, it all depends on the system for which you are developing the model. I am telling you even 0.8 also you can consider, absolutely no problem. If you think that the dynamics and is huge the volatility is more, many other issues you have to take into consideration. Now, another index is RMR, I think this is something where each of the value of the S matrix and each of the corresponding value of the estimated matrix, that sigma basically you take S then you estimate sigma and by that process in between the parameters are also estimated. Now, the values sigma and S values, S this values and sigma estimated, this values are here, getting me? Now, if you take this and this what is the difference? Take this, this what is the difference? So, we will take this, this what is the difference? These differences are squared here, you see what we have done S J K minus sigma J K this is the estimated one square by p into p plus 1, that is the non-redundant part. 2 is given because that twice of this, this by 2 p into p plus 1 by 2. So, this quantity should be also as low as possible. It varies from 0 to 1 and RMR should be less than 0.05 and for RMSEA root mean square error approximation, this is the modulus of chi square minus its degrees of freedom by n minus 1. It is seen that 0.03 to 0.08 in this range this lies and this range also says substantial increment then relative fit index. Now, how well does a particular model do in explaining a set of observed data compared with a range of other possible models. Here what you do? You creates nested model, several models and then you compare one model with other and then you say which model is better. And based on this you create a index and that index talks about your model adequacy or otherwise we can say improvement in terms of model adequacy. Here, most of mostly we will consider null model, that is the worst fit model which is known as null model, where we think that the covariance matrix is diagonal. Only diagonal means variance part is there of diagonal elements as 0. Now, if you say that X chi square for the null model is chi square 0 and chi square for the proposed model is chi square mu, where mu is the degrees of freedom. Then you are in a position to develop or I can say quantify, this indices like NFI is chi square 0 minus chi square nu by chi square 0 and all these indices this values lie between 0 to 1. And it is desirable that they will be greater than 0.09, sorry 0.90 then CFI comparative fit index, that is 1 minus chi square nu minus nu by chi square 0 minus nu 0. And TLI you have seen this, this also in the similar manner you see. Ultimately they take into consideration the chi square value of the proposed model and a worspit model which is known as null model and the comparative indices are developed and higher the index value, it is better 0.9 or more is required. Then your parsimony fit index, it is similar to adjust R square S A square in you regression and it basically talks about the par parameter fit. Par parameter is estimates and AGFI here you just see that the top upper portion or the denominator here is divided by the degrees of freedom and numerator is also divided by the degrees of freedom. What we have done in calculating R A square, this value should lay between 0 and 1 and AGFI greater than 0.90 is desirable. Then parsimonious non fit index which is nu by nu zero NFI, and where this is nu is the degree of freedom proposed model and like this, okay? Now, let us see the goodness of fit indices for the case study here, some of the fit indices I have given there are others. So, chi square degree of freedom of 99, chi square value is 257.24, if you divide by 99, this is almost 100. So, it will be around chi square by degree of freedom is around 2.6. So, it is good because aim is to do 5 root mean square residual 0.06 which is little more than 0.05, CFI is 0.98 very good more than 0.90, NFI 0.97 more then 0.90, CFI is 0.99 and IFI is 0.99. So, essentially then chi square is 257.24 chi square by degree of freedom is around 2.6, RMR is 0.06, GFI is 0.98 like this. This model is very good, fit model. Now, see that who has basically worked in this, who are the pioneers that Karl Joreskog and Dag Sorbom, I think in around 1978 probably they have developed this software. First that listenery came I think in 1989 and its remarkable development in this field and we all are tremendously benefited. What I can tell you further that for you have to understand the structural equation modelling. I might say that Joreskog, Sorbom this LISREL 8, structural equation medalling with SIMPLIS command language. This man, this manual is very good and you can go through and is a lot of publications by Joreskog, Soorbom. Others is the Hayduk is one person who has written a book, this and this is a very good book. Also in addition there are many other books available in structural equation modelling. So finally, let me just summarize to my today’s lecture. We said that structural equation modelling has two parts, one is measurement model measurement model, another one is structural model. So, measurement model is nothing but confirmatory factor analysis structural model is actually path analysis. Now, we have discussed details of CFA in terms of its identification, what I say that under identification there will be necessary condition. There will be sufficient condition, this two must be satisfied in necessary condition is known as order condition and this one is known as rank condition. Other one sufficient condition is known as rank condition and in this order condition we say the number of parameters to be estimated must be less than number of non-redundant elements in the covariance matrix. Then this is over identification case and it is a desirable case, then we have shown you the estimation parameter. Estimation, now I said that parameter estimation it is basically a function you minimize, which is basically log of determinant of sigma plus trace of S sigma inverse minus log of determinant of S minus p. This function is minimized through Newton Rapson or similar method and then the parameters are estimated and the actually we have sample data in terms of S or R. And we have the population value in terms of sigma theta. We try to match this two and using this function the better, the best match is considered and then you corresponding the theta value, these are used. Now, theta is function of many things like lambda, like your phi, like your theta delta. So, these are theta, means theta means so many things are there. Any combinations that is what you are trying to estimate because from co relation matrix to here, co relation matrix or covariance to co variance matrix, one to one, this correspondence you are doing. This is parameter estimation, once parameters are estimated then you can test the parameters values, this lambda using simple t test whether it is significant or not, but apart from this the another important output from this CFA is after parameter estimation is your co relation matrix of the latent correlation. Or covariance latent construct which is very, very important, because this will be the input to the structural model. Then what I have given you? I have given you the what are the model adequacy test. So, under model adequacy we have seen that absolute test and then your comparative test or relative test, another one is parsimonious, absolute is similar to R square R A square, where the actual variance explain is considered in comparative case. We compare with different model and parsimonious, it is basically fit part parameter estimated and finally I have shown you a case study for all of you. The case study is there, if you are interested please go through Paul P and J Maiti, the synergic role of socio technical and personal characteristics in mines, published in ergonomics in 2008. Thank you very much.
Info
Channel: nptelhrd
Views: 7,233
Rating: undefined out of 5
Keywords: SEM - Measurement Model
Id: JU-Y4utNxgY
Channel Id: undefined
Length: 60min 50sec (3650 seconds)
Published: Fri May 09 2014
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.