Design of Experiments - Overview

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
Good afternoon, we resume our lecture on Six Sigma and the topic that, I am going to be discussing today is a very key one in the process of achieving parts per million types of defects this technique, is called Design of Experiments and it is as I said it is integral of the six sigma methodology. Let us take a look at, why is it that we need to do experiment and what is it is role in the full range of quality engineering, which is actually today it has culminated into six sigma that is the state, where we are today. If you look at the picture today, you can see the clouds there and underneath is weather then of course, on this side we have got psychology and I have got the economy here doing its ups and downs, on this continue and if you come to the other extreme you got this digital watch, if I may show you my digital watch this is at other extreme of this knowledge that is there. On one end we have got very little knowledge about say something like the weather system, and on the other hand when I when I talk about digital watches, we are right there, we know 100 percent about how digital watch functions. In between our all the different system that we have around us medicine for example, chemistry, engineering, electronics on there all over the place they are on this spectrum. Now, what is the difference between systems on this side which is to my left, and the the systems which I have to my right hand side; it turns out that man has always tried to control systems, whatever we have around us, we always made an effort to try control it. And one of the things we have tried to do, is to try to first do basic research and of course, we began with observation and then it seem like some theory was behind it, then we put in our a lot of effort to try to understand the basic driving forces behind that process. In the process we ended up with some fantastic theories, the result is something like the digital watch or the computer and the today’s communication device and so on. Their knowledge has progressed to the point, we can write the exact equation how a digital watch functions for example, this is not so if you look at the weather system. The weather system has too many variables of you remember my cause and effect diagram there are so many factors, there are multiple factors that affect the weather system, and it turns out they are very very little is very known how the weather system behaves. So, we cannot certainly write down today’s, as up today’s state of knowledge we cannot write write down equations, exact equations to describe how temperature is going to vary, how humidity is going to vary and so on so forth, that we are not able to do today, as far as weather systems are concerned. Same thing goes for psychology, we are not able to write down the equations and also the same thing pretty well comes pretty close to economy when it come when it comes to talking about the economy, we know very little about the real process that is there and of course, theories are there to try to sort of you know and those are based on primarily observations. In between of course, there are all these other systems, but we people we men and women, we cannot really live with a system that is highly unpredictable, and we had the we are at the mercy of clouds and sunshine and so on so forth; we really cannot do anything about them, we really cannot with the system like that. So, there has been a constant effort to try to take systems which are on this slide, on this spectrum to try to gradually push them to the right hand side, so we can control it. This has been our continuous question, in fact it began probably 1000s of years ago and today, we have some knowledge in medicine, we have a fair amount of knowledge in chemistry, we have a lot of knowledge in engineering and certainly in areas like electronics we have really got all the equations known and determined and so on. So, basic result research and also applied research it has advanced to the point on the right hand side that we know everything about it, and we can write down the equation. Now, if you are trying to if we if our desire is to take systems that we know very little about and we want to push them to the right hand side, that we can control them, how do we go about doing it, well we can just stand around and observe that is like one way, just observe and try to see there is any kind of correlation between, what we see. And what happens perhaps there is some driving, some some cause in effects relationships that is how, perhaps these complex phenomenons take place for example, rain and so on and so forth. Then of course, the process can be further modified and this is where we have tried to use our intelligence, what we have said is suppose we are able to identify some factors and we are able to construct, what we what we call the fish pond of the cause and effect diagram; and then we manipulate some of these factors and see the effect. So, what we are really doing here is, we are intervening, we are intervening into the process, we first try to figure out what are the different factors that might be affecting a particular process, then we manipulate these factors in a manner that will produce hopefully some effect there, now this is the experimental approach this the empirical approach. You have manipulate some of the input conditions you take a look at the output, if can see some sort of dependency, there some sort of relationship then of course, you got some clue as to how this process this prestigious black box, how it operates, that is something that we can really figure out. If it, this is what is done a lot, if you look at engineering, if you look at chemistry, if you look at lot of these things, a lot of knowledge in this area, has been basically acquired through empirical work, through experimental work. Now, if we did experimental work that is fine, but it turns out many factors are affected by multiple factors, many systems are affected by multiple factors. That means; somehow I have to basically take four, five factors and I have to manipulate them together, in some complex way this is not a easy task, it is not very easy to manipulate four, five factors together then see the effect, then come back say come back and say that the response, a particular response was due to factor a, another response was due to factor b, another response was due to factor c. And perhaps something was also going on in the background when factor d was acting, this is not very easy to do, if you just do one factor at a time trials I will show that later; I will show why it so difficult to do one factor at a time experiments and hope to see that you get understanding of the process. So, what we needed here was a technique that could take multiple factors together and give us a procedure give us a methodology that could again give us insight as far the factor effects were concerned and indeed if there was any interaction between the effect, we should also be able to uncover that. This was done around world war two, and the gentleman who did this his name is Ronald Fisher, sir Ronald Fisher in UK, he basically come came up with the scheme and he did some mathematical work also, and he he was able to find a technique which is later on in mathematics it is called, in statistics it is called ANOVA, Analysis of Variants; this was the method by which you could find out, which of these four factors really has an effect, indeed if two of those factors interact if that is significant this is also you can find out. And also this led to slowly and slowly gradually getting to an equation an empirical equation form then of course, other people got involved and they tried to evolve the basic understanding of process later on as time went by. So, this technique this technique of manipulating multiple factors this came from this professor Ronald Fisher, we will see how this was done that is like the cracks of it. So, in fact what we are talking about is, we are talking about multiple factors manipulated together and some methodology that will let you see, the effect of each of those factors and perhaps their interaction also; this is the sophisticated empirical approach, this is very different and much more advanced when you compare that to one factor at a time trails, it is quite different from that, let us see how you get into this. If you look at any of the today’s systems if you look at any of the today’s systems, you do not have to really look at the digital watch that is far to complicate, if you look at of factory for example or a machine you know something that is doing grinding or something that is doing welding and so on so forth. If you look at any of these systems generally speaking, what you will find as you see see on the slide, there are multiple factors that might be effecting the process, some of these these we call control variables and these are of course, on the input side I have got raw material, I have got labour, I have got training level and so on so forth. Then I have got process process conditions, various process conditions which are there, these are going to be influenced by the control variables, these are going to be factors that might be in impacting the process that is inside. In addition of course, we have also got ambient temperature vibration humidity and so on so forth, these are external conditions and these conditions are not necessarily always in our control, and somehow we have run the process in presence of all these noise factors. Anything, that you are not controlling we regard we the professional we regard those as noise factors, so me how what I have to do is I have to figure out, what is going on inside this black box which is the process. If I had the equation of course, there will not be much difficulty in coming out and writing down the equation, the exact equation that will show the output as a function of the input factors. When that equation is not available I basically have to do empirical work, and this is done a lot in R and D labs, in the R and D labs they manipulate these different factors try to and they try to see, what is the effect of it. Let us see how we do that, the beginning of course, is again this fish pond diagram where I have got some speculation of how a particular process works and I have listed out the factors here, and the output is on the right hand side which is effect of all of these. Now, if I have to study these if I have I study these different factors I certainly need some special methodology, I cannot just do that the way ordinarily experiments are done, I cannot do that by manipulating one factor at a time just cannot do that, I have to manipulate multiple factors together. Because, the real process has the involvement of all these different factors there therefore, what we have to do is we start with the cause and effect diagram and then we come up with a procedure which will allow me to set the different settings and these are called treatments. Of these different factors which are there, then look at the output and then do some special data analysis which will let me see, whether factor A was important it had an effect strong effect on the output or B was important or C was important or D was important or was there an interaction between A and B or A and C and so on so forth. Now, the the process the experimental process involves all the factors together I am not doing a one factor at a time trail that is not the way I can move from left to the right, I cannot go from the state of zero knowledge to perfect knowledge, I cannot do that by manipulating one factor at time. Partly because, factor effects many times they interact and that is a very complex phenomenon and that is why I have got to manipulate multiple factors together, and this I precisely what is done in DOE, Design of Experiments is the special method that Ronald Fisher came up with. So, if you look at a process, I have process parameters inside and I have got all these different factors and I see this cause and effect dependency there, the effect is the output and the causes are any of these factors which are to the left of this box there. And my goal is going to be to manipulate these factors in some systematic way, so that I can work out the dependency of the output to this input conditions; this is basically what we are trying to do, what we would like to able to do. Now, let me tell you in a snapshot, let me tell you how this session is going to go and how the other sessions and design of it is, how they are going to go say you got three factors, you have three factors and notice a box there, notice a box to the left this is actually 3 D matrix, that is all it is. Now, when I am trying, when I am speculating about the effect of A, and the effect B, and the effect of C, what I do is I construct this matrix this matrix is an arrangement that led lets me manipulate simultaneously all three factors; in a particular way, in a special way and let me show you, let me actually show you, how this manipulation is done. We will assume for a minute that factor A has two alternative positions, factor A can be set at low level and also it can be set at a high level, so in fact factor A has only two levels, for simplicity I am assuming that; factor B also let us say has two levels low and high, factor C also has two levels low and high. If I combine two times, two times, two I have got eight combinations, so there are eight different ways in which I can run this process, when I am manipulating these factors A, B and C each can be set at two levels each can be set at two levels, what I am really trying now is, I am trying to sort of see, can I combine two and then two and then two, can I combine them and that combination is done by this special matrix there. And I am going to tell you a lot more about these matrices, these are called DOE matrices, these are special factor manipulation plans, these are special plans that allow you, that really enable you to conduct these different experiments appropriately adjusting the different treatments that this process is subjected to. What is the effect of all this in between you see that fat arrow there, transfer the data comes out, I run each of these experiment, so for example, this experiment here will be run at the setting up A at high, setting up B at low and setting up C also at low, so I have got high, low, low, that is going to be the setting for this experiment. It will have three factors set at high, low, low level, when I conduct the experiment, when I run the process by holding A at high, B at low, and C at low, I get some response, I record that response, that is like some of the outputs from my first round of experiment. My second round could be this one, when I got A at high level, B at high level, and also C at high level; so the setting for this experiment setting for running this experiment is going to be high high high, I do the same thing for each of these corners there. And therefore, I end up running eight different rounds of the experiments, I have now run, because of this scheme there, because of this matrix scheme there I have run eight experiments. So, the data that I have now on this big fat arrow there, this data now has the results of eight experiments, then there is a data analysis procedure, I apply some data analysis procedure to the data that comes out and the result of that is going to be, I will be able to discover calculate actually main effects these are going to be the single factor effect, effect of factor A, effect of factor B, effect of factor C. This I will be able find out individually, even if I would run the eight experiments by manipulating all three factors together, the data analysis procedure is so good, that it gives me separately the effect of A, the effect of B, the effect of C that I am able to see, these individual factor effects these are called the main effect. Now of course, there are other complications where the effect of factor A and factor B they may interact, this also you can find out from the same from the same data, the data that you have there from the eight trials, you will be able to find two factor interactions and perhaps even three factor interaction, because I have got A, B and C three factors, they may interact in a complex way, that also I will be able to uncover by doing this special data analysis technique. So, in fact the reason is, these plans are special plans, these plan plans are actually very special plans, the scheme is a very special scheme that scheme is unique and that will allow me to separate the effect of A, effect B, and effect of C, and also the interaction of the two factors two factor two factor interactions, also one, three factor interaction. Because, there are three factors they will be one in fact it will be it will be just a setting of certain values of A and B and C and I will be see, the interaction effect of all three working together. Now of course, there is some other line written there notice here, there is I have got a statement there called significance, I have got a phase there called significance, I have got main effects interaction and significance significance is a statistical significance, it is a statistical term that actually says the effects that I notice, they are going to be significant they are going to audible in the background of noise, it is sort of like this. Let us say, if this studio had a lot of a a it had a noisy fan running and I was trying to do my lecturing in the presence of that noisy or the blower that was there and I can simulate that by saying somehow I will raise the noise level and either, I can raise the noise level and let you see what happens or I could reduce my voice to the extent that it becomes almost indistinguishable to the background noise that is there, and let me just try that. So, if you notice now my voice got lost, the effect of my voice chord and breathing and everything it got lost in the background of noise, in fact if someone was sitting next to me even next to me he would have a pretty tough time to find out what I was saying, when I talk about factor effects. Each of these factor effects are significant relative to noise, relative to background noise, because that background noise will be there, but what statisticians have done it, they do not say that the effect is significant unless it is like a multiple you know significantly higher than what background, whatever that background level of noises; that is actually how they figure out when it the effective significant, I need to worry about it or I can just let it go without doing anything. Now, after I have got these effects calculated I can do some plotting and notice here I have got the effect of A plotted out, this like when A goes from low to high setting, this is what happens to the response, the response goes up when I move the setting of B from low to high the response comes down. And when I manipulate C from low to high again the response comes down, out of these factors A, B and C you can clearly see, that C has the largest effect, C has the strongest effect; and if my goal was to try to maximize response, I would have set A at high level, B at low level, and C at low level. If I would pick this setting for A and this setting for B, and this setting for C the combine effect hopefully is going to be maximum, the maximum value for the response that is like an empirical optimization that manipulation actually has been taken place empirically. What else can I do with the data that I have there, there are techniques available with the help of which, I can take the data that came out of that green fat arrow, I can take the data and I can come back with an equation and look at the equation there Y is the response, I have got alpha plus beta times x 1 plus you know chi times x 2 plus delta times x 1, x 2 and so on so forth. So, this is now a mathematical model and I can construct this model by regression analysis using the same data that I have got over here, I can use the same data I can bring it through and I can do multiple regressions and I can find out this equation, I can find this equation. Now, because this equation has been derived using empirical work, we call this an empirical model, what the use this empirical model, remember the ultimate goal is for us to get in there and to make everything controllable like that digital watch, now fortunately for the digital watch, we did not have to conduct experiments. Here, we could study ohms law, Kirchhoff law and many other things, that that was provided to us by people from physics the Ph.D from physics, they gave us the equations, and they gave us those electronic equations and so on so forth; and we were able to design this watch based on those equations. In this instance, when I am trying to manipulate factors empirically I do not have those equations, but I can come up with the proxy equations and that proxy equation is this regression equation, which you see at the bottom of the slide here, Y equal to alpha plus beta times x 1 plus chi times x 2 plus delta times this things, this equation is the empirical equation, this empirical equation is almost as good as that original theoretical equation, if one able if one was able to construct it. What is the use of this, well I could go to the next state, I can optimize this function I can find the values of x 1, x 2 and so on so forth; that will maximize Y or minimize Y whatever be the be the desire that optimization, I can now do using this empirical equation. So, just see where we started from and where we are, we started with the cause and effect diagram which is like this one, and I conceptualize them identified some of the factors and I used them in my experiments, and I constructed this matrix this is a special matrix and after I ran the experiments collected the data figured out the main main effects and the interaction effects, constructed the regression model and then of course, I apply optimization. Now, what do you do, what what do you really try to optimize well, if it is like a product system and we try to maximize productivity, if it is a scheduling system I may try to minimize, the time to finish all the jobs, if it is a chemical process I may try to maximize utilization of raw materials, I may try to maximize the utilization of anything else that is there or I may try to maximize yield these things are possible when I apply optimization. And this is quite easy when you got equation in front of you, whether it is based on basic theory or it is derived by empirical work like we done, in a lot of work in six sigma, we would be forced to derived actually these, construct these empirical models; and therefore, six sigma experiments also are going to be special experiments, they are all going to be guided by DOE the general principle of DOE, what is the effort, what is actually the aim of six sigma, six sigma tries to reduce variation in quality. If you got any kind of variation in quality for example, if you look at the output of this process here, the output is highly variable and this may not be satisfactory for a customer, what we would like do is, we like to make this process as tight as we can make them and once we make the process tight it is going to satisfy the maximum number of customers. So, our goal is going to be then to try to shrink the spread of the process and I will show you later how we end up doing that, this is this is to be by us we are the specialists, we are going to be doing conduct some experiments, and if you follow the domain procedure of six sigma, define measure, analyze, improve and control, if you follow these this five step procedure, at the improve stage we are actually using DOE. And that is when we are trying to manipulate the different variables, find the effects and also optimize the process all those things could be done, with the sole goal to try to minimize minimize variation in the process in the output of the process; that is like one goal also in the process we will end up producing defect that is like also another objective of conducting experiments. There are many techniques available of course, many statistical techniques available and you already seen histograms, pareto charts, fish pond diagram you have seen tabulation of data you have seen run charts, scatter plots, flow charts, these are techniques that almost anyone can use. In fact, the the work force should be trained on these techniques for sure then of course, the engineering types of the production supervisor they should be able control the process online and they can use control charts. So, training also should be provided at this level, so that they can manipulate the control charts, they can actually use the control chart to try to locate any kind of defect that is there for which there is an assignable cause. Then of course, we got these techniques and these are the most advanced statistical techniques, these are part and partial of six sigma black belt black belt training and the black belt people they can teach this to others also. What are the techniques involved here, we already heard about DOE, we also heard about regression models, so there are certain other things called significance tests and confidence interval, these are advanced techniques which are all used if you are trying to optimize the process; and six sigma uses this the top of the pyramid six sigma very heavily uses these techniques in trying to do it. So, every black belt should know techniques from top to bottom, every black belt should really know all the techniques that are there and specially they should really be focusing on this and they should be able to help, other people who are trying to use DOE or regression. What exactly is an experiment, I gave you an overview, now what exactly is an experiment, is the process of systematically varying the variables of interest, these are the control factors about which you got some speculations, and the idea is to try to make sure, you are able to see the effect and you are able to you are able to do this with by manipulating some of the input conditions and by observing the output; that is an experiment, we do experiments all the time. In fact, it is surprising how often we are conducting experiments and we are very mindful of it, we are actually indeed we learn so many things by doing hit and trial, that is also an experiment, we make some change in the input, we change the input and we look at the output, and we try to see is there a dependency there, if there is something then we learn something, we find out how this this particular system works. Every time we are doing a trial, we are measuring response, every time we are doing a trial in our experiment, we are measuring the response. And some of the critical concepts which are involved in an experiments are variables are manipulated by the experimenter, random assignment of experimental units are there many many times what happens is, if you let us see you are conducting an experiment and you did some of the experiments early in the morning when temperature was low. And the sun was not there, and then you conducted some other experiments at lunch time when temperature was thigh and the sun was out and the sun was pretty bright and so on, and then you conduct the same kinds of experiments you again conduct in the evening. Now, you are manipulating only the variables in which you are interested, so you went to your fish pond diagram and you picked up your control variables from there and those are the ones you manipulate using the matrix that is there. Unfortunately while you are doing this, sun is also doing its own things, so it was cool in the morning it became warm then perhaps it cooled down again, that also might have an effect on the experiment, therefore, what you have to do is you have to randomize you have to randomize the order in which you run your trials, this tends to fool factors like sunshine and so on. You will end up finding some effect of that, but it will get randomized they found the average there those things are going to cancel out, and this I am going to discuss again as we get deeper into this we are going to be doing this. And something else, we should also try to do and this is kind of a somewhat of an advance concept, we should try to avoid confounding of variables that means, in places we are manipulating one particular variable, another factor should not be confounding the same the same effect it should not actually be such. Then I am trying for example, to find the effect of one factor let us say that is my left hand and I am trying to do something with my left hand at the same time I should not also bring my right hand and try to do this, because then when the effect takes place I would not know is it due to my left hand, is it due to my right hand, here the effect of the my left hand and my right hand they can found it. This makes it very difficult to figure out which factor actually was it my left hand or the right hand that produced the effect, the same the same idea we use also when we are designing experiments, we try to avoid confounding we let the variables stand on its own whenever we do settings of an experiment, we make sure if we are if you are manipulating a particular variable in a certain order. There should not be another factor that is also manipulating in the same order, if you did that if you did for example, temperature high and low and also the same setting we did the pressure high and low simultaneously we did not run any other trial, it will be very difficult to figure out what is what the the effect that is observed is that the due to the effect of temperature or is this the effect due to pressure, is effect due to pressure. Now, if you conduct if you conduct some experiments and some very simple one of course, are there for example, this is of course, not an experiment, but you pretend that some factor is being manipulated and you are looking at the response so for example, number of packs of a cigarette smoke per day, so half of pack or one pack per day, one and a half pack per day, two packs per day, two and a half packs per day. Now, something that is really correlated is well established a medicine is the severity of cancer, severity of people having cancer on some scale that is got a correlation with the number of packs of cigarette smoked, you see that raising trend there that raising trend actually says that there is a relationship between, there is a dependency between number of packs of cigarette smoke per day and the the severity of cancer that the man, that the person contracts. So, this is actually is something fairly simple and easy to see, when I have got only one factor, notice something else there are these, this not quite a straight curve, this not quite a straight line or a smooth curve there are some fluctuation, there are some random variations there and that is because, at some point in time the total effect of causing cancer that is not the effect of smoking only, it is got genetic orientation and lot of other factors which are there, that will differ people to people it may differ from city to city or situation to situation those also have their effect and those those who actually make people cancer prone. Let us see, if I will if I will living next to a refinery and the refinery is being out all this sulfide dioxide and nitrous oxide and so on so forth, those are also have an impact on the person therefore, it is not just going to be his smoking habit, it is also going to be effect of this environment that is going to let him be kind of more like you to getting to cause to catch cancer. What we have to do here is therefore, we got to worry about combining not just using, just not one factor in my experiment, but also incorporating other factors also, these are those are that also speculated to be effecting the process, that also would be there. So, correlation is something that is quite easy to see for us for example, we see correlation here and this is just defined here, that correlation is something that we find to occur together, find to be occurring together for example, number of packs of cigarette smoke versus the extent of cancer contraction, the these are like these seen to correlated. But, of course, something I should caution you about just simply observing a correlation does not really mean that one is causing the other, just simply because, I have got some data and plotted the data and I see this curvy kind of thing going up and I see then kind of going up together. This by itself is not the proof, that the cause is affecting the effect and a cause has anything do with the output correlation, the correlation relationship is not really guarantee that one is causing. Therefore, what you have to do is when you manipulate the input, we then have to observe output and we have to really see then, after manipulate the input variable does that produce a corresponding change in the output, this is an experiment; this is not passive observation this is not passive observation, here we are deliberately manipulating a factor an experimental factor. And then corresponding to each of those setting observing the output, then if I see a dependency, if I see a dependency then then I will say perhaps this is causing that, and this again I have to see in the background or noise that is there, because noise is also affecting the output therefore, what we have say is, it is just like my voice if my voice was so low, and the noise was so high in the room that you could not hear my voice, perhaps and what you to do is you have make sure this signal is stronger than the noise, only then you say that the signal really is significant otherwise it is not. So, correlation by itself is not a not a sign for causality for that you have to conduct some special experiments, what kind of factors are we talking about of course, there is something called the independent variables these are factors that we manipulate in an experiment, and the dependent variable is the response, control variables are those those things that you might like to hold constant when a process is running. And random variables are those factors that you are not controlling and no one is manipulating them and they are changing on their own for example, vibration and voltage fluctuation and so on; those might be changing on their own for example, and of course, something that is called a confounding variable is I have got one factor already set at a certain level in a particular experiment. And I am putting another factor, in fact coinciding with the position of the first factor, we should have to we should try to avoid such settings and experiments, because then it will be very difficult for us to say, whether the effect was due to my left hand or was it due to the right hand, this is something we should try to, try not to get into, if we do that our experiment design will not be a satisfactory design. What is a noisy experiment, on the x axis I have got different experimental conditions a person is given no drugs or he is given just sugar pills or he is given cocaine in an experimental setup, on the y axis I have got his heart rate I will measure his heart heart rate and notice here there are these dark green box and there are these dark green blocks and each of those settings there but, also there are these light green box what I really observe is not only the effect of the this dark box, but I observe this one. So, in fact, when I am watching from this side, I will say 90 for this response, I will say may be 80 for this response, and probably 95 for that response there, now what is the problem here, my experimental factors are these drug, no drug for example, or sugar pill or cocaine, these are my experimental factor setting unfortunately, while these experiments are going on the background is highly noisy. For example, if it is a matter of heart rate perhaps, the person when he takes these medicines these conditions, he just see these things perhaps at the same time, he is also doing a treadmill run for example, or he is jogging outside is something like that, I have given him this pill and he has gone out he started jogging. Now, the effect on his heart rate is going to be totally due to this pill that I have given him, if at all there is an effect of it, but also due to jogging or due to his kind of running on a treadmill for example, such a background is a noisy background, if I am trying to see an effect and that say effect is produced possibly by a noise factor, I am not I have not designed my experimental conditions very well. What is the choice, we will just control noise look at this condition, look at this experimental, setup this is a well designed setup, what what have we got there I have got here no drug situation sugar pill and cocaine, but I have controlled background noise very well, the result is that I am able to see indeed cocaine causes the maximum rise in heart rate, sugar pill does not do that much may be it does just a slight bit and no drugs of course, has hardly anything, it is just the resting time, whatever it is. So, this is now, this is not a noisy background and I am able to see that, the lesson from this is when you are trying to conduct the experiments, do not run the experiments under these conditions, try to raise your experimental conditions in such a way, that you end up with results that are like this; except in a case and except and in a very special case when you want noise also part of your experimental study. And I am going to give you an example of that, when I tell you about the design of chocolate bars, chocolate bars are pretty fancy things to design and when I design the you know when I when I talk about the design of chocolate bars, you will actually see sometimes you would like to have noise also included, when I am doing a design study, when I am doing for example, some experiments I may include noise, that we will come back to in a little while. So, in fact did the what we have what we have figured out, so far is we have got multi factorial situations that we must manipulate and typically we will be manipulating more than one variable, that is what we will be doing, we may change more than one factor, we may have their distinct settings there for data analysis we will be using an anova that is something that we would be doing; our ultimate goal is going to be to try to construct some sort of a response surface, so that we can come back and find the optimum settings for the experiment, without doing that we of course, cannot really say that we have really help the process. Let us take an example, let us take a very simple example, this is an example that is a six sigma case and the problem was this that many customers who bought Nissan cars, they found after a while that the logo fell off and you notice their logo there, the logo is right there, that logo is stuck with some pressure and some glue the logo is stuck on the back of the car. This little styrene there, it is stuck right there and this too go would keep falling off and this was a kind of a somewhat have a mysterious situation, people did not quite know, what was going on and they did want to kind of you know do something about it. Now, one way would be to be just basically try different glues or different weights of logos or different sizes of logos, and it basically manipulate only one factor at a time, you could do that, but would that really work suppose I change the logo and I also change glue. What would happen, for this I needed to manipulate 2 or 3 or 4, 5 factors together and that is shown in the matrix which is here, this is the matrix that actually gives you an idea of the different factors that might be affecting that logo there. So, let us just for a minute let us draw a little cause and effect diagram, I have on this plate here, I have got logo falling off, logo logo falls off that is the effect, and what are the different factors that might be causing this, I am again going to speculate I do not for sure it could be glue, it could be the size of the logo, it could be the thickness of the logo, it could be the pressure that I apply, any of these things are there may be many other factors also those might be effecting this thing. If I look at this situation since any of them, any of these and perhaps their interaction may be causing this problem there, what we have to really do is we have to conduct an experiment and for that we come back to the slide again and notice here these variables adhesion area, the area over which you apply the glue, type of glue that I use, thickness of the foam styrene which is their between the car and the logo, because the logo is a metal you do not want to damage the car. Thickness of the logo itself, amount of pressure that I apply in trying to make sure, I stick the logo firmly enough and I let it dry, let it set then I remove my hand there. The amount of pressure that I apply pressure application time, that I just mentioned and any on a primer that you apply on the surface of the shining car which is there, that shine itself that that polymer that is there that itself, might resist sort of proper gluing. So, that also is something that I might like to manipulate, now I would like to run a simple experiment and what is the kind of experiment that I am going to run, I am going to pretend that there are two areas available, 15 centimeter square and 20 centimeter square type of glue acrylic glue and urethane glue, these two types of glues are available to me thickness of the foam it could be thick or thin, the thickness of the logo it also could be thick or thin, amount of pressure could be short or long small and big so on and so forth. Notice each of these factors have now, what we call treatments, two treatments for each of these factors, so if you now look at the look at the experiments that were run, what experts suggested that instead of trying all seven factors, we should focus perhaps on the first four and this seem to be the ones, that could be the culprit that could really fix the problem. So, for that, the first four factors were chosen adhesion, glue type thickness of the foam and thickness of the logo, having done this. These setting too were taken out and they were put into matrix and notice the matrix there, this is the matrix this is the matrix, this matrix has these factors there factor A, factor B, factor C, and factor D and those are written up here, A represents adhesion area, B represents glue, C represents the thickness of the styrene foam, styrene foam that is there and D represents the thickness of the logo and that is what I have got allotted there, I have got column A, I got here a there A sign, then B A sign on the third column and then I have got C a sign and D a sign. Now, the first trial that I will be running, I will be probably be mounting may be ten of these logos under conditions which is like A at plus settings which is the high setting, B at plus setting that is a high setting, for a C at plus setting and D at minus setting. And I run this trial and I measure the gluing strength, then measure this strength with which I can now remove that that that that strength, that is there I have to apply some force to take it off, I do the curing and then I try to take this thing off, and I measure the strength that is required force that is required in trying to take the logo off; that would have something to do with the with the lasting quality of that logo sticking on the back of the car; I do the same thing there, so gluing strength is actually the response and these are the manipulated variables. In my second trial I do A at high level, B at high level, C at low level, D at low level and so on, and so forth. So, I run eight different trails I run eight different trails and observe the growing strength for, each of these things I have got for each trail, I have measured my glue strength notice here a couple of things this is a planned experiment, I have four factors that I am playing with, I have two settings for each factor. So, I have got 2 to the power 4 in fact, I have got now 2 times to which 4, times 2 which 8, now in fact I have got eight trials here, but I am manipulating four factors and I have done a very special construct of the matrix, this matrix has been specially constructed, I will show you later, how you pick the right kind of matrix to able to do it. Once I have done that, I can run my trials and I made my observation and this is like the the green fat arrow that was there, that was bring in data, so I have got my data now which is on the, in the last column; my data is right here, my data is in the last column there. I take this data and I do some calculation, those calculations basically we show me the effect of changing A from low sitting to high sitting, and look at the impact glue changes, I do the same thing for B, I do the same thing for C, I do the same thing for D, these calculations had been done in a special way, by starting with these input conditions and by looking at the construction of the matrix, the structure of the matrix; once I have done that I am able to calculate, I am able to generally I am able to calculate these factor effects there. Now, what this is showing is, when I change from plus to minus these are like two settings for a glue instant changes, it has the same impact on glue instant. It may plot it up the plotting goes like this A is going from you know high to low in this particular case and the effect is that my gluing strength goes up, when I change B this is the effect, when I change C this is the effect, when I change D this is the effect. Now, by doing those eight trial I conducted exactly eight trials, and eight is like this, with those eight trials I have been able to study four factors and each of those four factors, they were set at two levels and my data manipulation technique by data calculation technique. Let to these plots let to these plots, that you see there what does this tell us while as far as this gluing process is concerned, additional area is the is the one that is really the culprit, perhaps we were not using sufficient additional area there other factors seem to have an effect that is more less flat; so I need not really worry much about manipulating those of the factors there. If you really want to we could actually take them to the level where they give us, they they would give us a high strength that we could do, but as far as fixing the process is concerned, if you would change the setting of A from minus to plus immediately give us a big jump in gluing strength. This came out as the result of doing the experiment and the red circles they show us the optimum settings now, for basically running the process basically running the process which is there. What are the steps in conducting the experiment just take a look at, how we conducted the experiment, we stated your objective then we also selected the response, then we selected the factors of you wanted to manipulate, we choose the factor levels for each of these factors the x factors, we selected the experimental design which is the matrix I have been talking about, then we ran the experiments in a randomized order and we collected the data, then we analyze the data and we get got those plots done and we do some conclusion. And of course, once I come up with a with an empirical conclusion, I still have to conduct what we call the confirmation run, I have to convert one conduct, one experiment under optimum condition. So, for example, for the gluing experiment I should run one experiment that sets A at minus, B at minus C at plus and D at plus, I must run one trail to basically confirm what we have deducted what we have deducted basically, what reduce empirically, that is something that got to be able to do. What are the advantages of doing it this way a lot of advantages, I could produce a lot of information by the conducting experiment, and this is like something that I was able to do without too much trouble. I could obtain the main main effects, I could also work out the secondary effect if there is any secondary effect, I could also work out at also calculate the introduction effect; now this could be done for each of the factors that are been manipulated, and in fact it turns out, because of matrix are very simple to use and the settings are all clearly laid out in the matrix. So, anyone who really understands how to manipulate the variables, you can take one of those matrices and run the trial, so running the trial does not really require you to know lot of statistics that is something very easy in modern experiments. In the older experiments, you also had to have to do something called anova, and its variance to come up with with your optimum settings and so on, so forth. Now, you do not have to do that, you do not have to really worry too much about significance, if you control the background noise, and if you use a simple matrix you should be able to work out the factor effects and their interaction that you should be able to do. And this is far more useful than one factor at a time experiment, which we are going to see again as we go later on, later on as we go deeper into this topic, you will be able to see that DOE produces information that is substantially better then the inversion that are produced one factor at a time, for one thing something that the one factor time experiment will not be able to able to tell you is whether there are effects between, whether there are interaction between two factors that can be manipulating independently. That effects point interacting some complex way, and this would not come out if I would come out one factor a time experiment that is like something that we got be able to keep in mind. And DOE is also pro active too this is something that I could in R and D, I need not really go out and wait for production to began and the real plant is running there and I start my experiments, my my development work there, that is actually not necessary, that is actually not necessary, I could basically run all these experiments right at the R and D. And really they would not be much trouble at all, so other advantages are also there for doing DOE, it can identify the critical precession parameters to control a process and to improve it. In fact, we saw that factor A was the most significant one the others did not really matter, as far as fixing that the dots and problem was concerned, that we could see quite easily. New product development when you manipulating many different variables DOE is a is a short cut way, DOE is an efficient way and it identifies the correct levels to help maximize performance and also to reduce defects and the overall cost, that also can be done if I set up those as a objectives of the experiment, I am able to those quite easily. DOE does something even more DOE does robust design, DOE helps us to robust design and this is the method that was used by Taguchi and that connection again I am going bring back the chocolate bars, I am going to tell you exactly how to construct robust design, that is robust chocolate. And that property I am going to keep a little secret right now, till I start discussing Taguchi methods that I will do, and the factorial experiments they are turn out to be most economical, and they are also very precise. So, there are lot of advantages there and there are of course, there many different techniques and one of the popular techniques for doing experiments is to use the full factorial design, this is a most complete technique, it can find a main effects and also it can find interactions, so both of those things can be found if you use full factorial designs. All combinations of the different treatment levels those can those can be powered by doing the full factorial trials that is there but, if these experiments become too many I have to use some other method which is called running a fractional factorial run, and that I will that I will show you once we get into that, I will show how we come back from full factorial design to partial factorial design. That is actually pretty efficient way to find a lot of information without having to run all the different to the 2 2 to the power k experiments, which is like the full factorial design, we will continue with our lecture in just a little a while, thank you very much.
Info
Channel: nptelhrd
Views: 75,453
Rating: undefined out of 5
Keywords: Design, of, Experiments, Overview
Id: sIRl1xWrViY
Channel Id: undefined
Length: 54min 16sec (3256 seconds)
Published: Mon Jan 02 2012
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.