The END of Photography - Use AI to Make Your Own Studio Photos, FREE Via DreamBooth Training

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
Greetings everyone. In this video, I will show you how to install and use DreamBooth extension of Automatic1111 web UI from scratch. How to generate amazing quality images like this. On the left, it is my real image. On the right, we are seeing the raw output of my trained DreamBooth model. Here again, on the left, we have my real image. On the right, we have the raw output of DreamBooth model. And here on the left, my real image and on the right, this is inpainted output of the DreamBooth model. I will show all of them. By following this tutorial, you will be able to generate studio quality images that you can use on your LinkedIn profile. You can use on your Twitter profile. You can use on your GitHub profile. You can possibly use on your official documents or wherever you need studio shot quality images. This will save tons of time of yours. This will save you huge amount of money. So this tutorial will teach you how to obtain amazing realism. So I have prepared an amazing GitHub readme file for this tutorial. Every information, links and commands will be in this file. This file will be the primary source. As it be necessary, I will update this readme file. The link for this readme file will be in the description of the video and also in the comment section of the video. When you click here, you will get to my Stable Diffusion repository. I am posting a lot of useful stuff here. So please star it, fork it and watch it and check out the links and descriptions that I put here. So we will begin with installing the DreamBooth extension of Automatic1111 web UI. If you don't know what is DreamBooth training and if you wonder and if you want to learn more details, watch this tutorial that I have put the link here. To be able to follow this tutorial, you need to have Python and Git installed. If you don't know how to install Python and Git, look out for this tutorial video that I linked here. You can download the Python 3.10.9 version from here and Git from here. If you don't know how to install Automatic1111 Stable Diffusion web UI, I made a recent video and in the first 15 minutes, I have explained how to install it very detailed. So watch this tutorial for that. We will use realistic vision version 2 for this tutorial as a base training model. The link is here to download it. So I will show both automatic installation and manual installation. I have prepared an auto-installer script. When you open this link, you will get to this: Patreon post that I have downloaded part 1 and part 2.bat files. Don't worry, I will also show manual installation as well step by step. So let's make a new folder in my c drive. DreamBooth tutorial. Let's make two folders, auto, manual. For auto installation, copy paste the downloaded bat files here. First execute part 1. It will ask you to run it. Click more info, run anyway. You can right click the bat file, edit and see what is inside it like this. So here I have my manual installation of Stable Diffusion web UI. Let's say you want to use the exact version that I have used for this tutorial. Then after cloning your Automatic1111 web UI, execute this git checkout command. Open a cmd window inside here, copy paste it, and you will get to this checkpoint. Then do a fresh installation. So download the realistic vision model from this link. When you click it, you will see it will start downloading like this. Put the downloaded realistic vision model into your models folder. Don't forget to join our Discord channel for any kind of help and also discussions with other expert people we have. Click this link. From here join our server. Automatic installation part 1 has been completed. I closed the cmd window. Then start part 2. Click more info, run anyway. When you see the url like this, it means that the first part of the automatic installation has been completed. And now automatic installation will automatically clone the DreamBooth extension and install everything automatically for you. Manual installation is also ready. So let's say you already have an existing Automatic1111 web UI instance. You don't have to do a fresh installation. All you need to do is start a cmd window, do git pull which will update the Automatic1111 web UI to the latest version. If you get a message like this, then do git checkout master and then do git pull again. And it will update everything to the latest. Let's say you have previously installed DreamBooth extension. Then go to the extensions folder and delete it. This is important. You have to be inside this folder. Open a new cmd window. Copy the command written on the readme file here. Paste it and hit enter and it will clone it. Let's say you want to use the version I used in this tutorial. Then what you need to do is executing this command, copy it, move into sd DreamBooth extension folder like this, then execute here. And then you will use the same version that I used for this tutorial. Why this is important? Because sometimes the extensions may get broken temporarily and when you are watching this tutorial, if it doesn't work, you can use the exact versions that I have used. Then close this window, move into your main installation folder, go to virtual environment folder venv, go to inside scripts folder, open a new cmd window here, copy this, activate and type it and it will be activated. You see now the virtual environment is activated. I have explained a lot of details about virtual environments in this tutorial video. Then execute these commands one by one. Copy this, paste, hit enter, paste, hit enter, then copy this, paste, hit enter, and then copy this, and paste, hit enter. And it will install all of the requirements of the DreamBooth extension. This is the most proper way of installing DreamBooth extension. Do not install it from extensions tab of the Automatic1111 web UI. Close your Automatic1111 web UI and follow the steps that I have shown here. So what is the benefit of using my automatic installer? If there be an error. Just leave a comment here and I will update the scripts and you will get the most up-to-date fixed script file and you won't have any issues. Alternatively, leave a comment to this video and hopefully I will update this readme file and fix the issue for you. So now we can use both automatic and manual installation. I will continue with automatic Stable Diffusion and DreamBooth installation. I will start my web UI instance. The web UI has started. You see the commit hash and the DreamBooth hash revisions here. When you open your interface, you will see the DreamBooth tab. So let's begin with creating our model. Click the create tab, give a name, any name. This doesn't matter. Me Best realism: select your source checkpoint. Click refresh icon here If you don't see the model. Select realistic vision. This is for realism. Check unfreeze model. You may wonder why check this? because this is slightly improving the quality. I have done over 30 different experimentation to find the best parameters. Hit create model: follow what is happening in your cmd window. This is important. Also, in the very bottom of my interface you see it displays version of the Automatic1111 web UI, python, torch, and xFormers. Currently my xFormers version is the latest version. Why? Because the automatic script automatically upgrades the xFormers to the latest version. If you don't know how to upgrade it to latest version, watch this tutorial. If you don't see any progress in your cmd window, hit enter because it means that it was frozen and then it will continue to display messages and continue operation. Okay checkpoint successfully extracted. You can see where the extracted model files are stored. It is inside in this folder. Let me open it. So here the extracted model files. This is where Stable Diffusion training model files are saved in this folder. You can see with the name of your training model and other variables are her. In this folder you will find the native diffuser files. So based on these diffuser files it will continue do training, continue doing fine tuning. You see all of the files are here. Okay, then select your model. When you first time create your model, it will automatically select the freshly composed model like this. Later time if you want to continue then you need to select your model from here. It shows the loaded model, model revision, model epoch, and the source checkpoint used to compose this model. I will show the best settings for the best possible realism. This is usually best settings for fine tuning and other tasks as well. So what matters for best realism is the used base model, used resolution and used images quality. I will explain all of them. First click performance wizard then click train wizard person. I train up to 200 epochs. Save model frequency is 10. So with each 10 epochs we will save our model. I am not saving any previews anymore, it is just slowing us down. I use x/y/z plot to find the best checkpoint. I will show you how to do that. For training the best possible realism I suggest you to use batch size 1 and gradient accumulation steps 1. Class batch size doesn't matter because you should never use DreamBooth extension to generate your classification images. Either you should use pre-made classification images or use text to image to generate your classification images. I don't check gradient checkpointing because I have 24 gigabyte VRAM having gpu, but I will also show the best settings for 12 gigabyte having gpus as well. So the learning rate will be 1e-7 this is important. Make the learning rate 1e-7 both for UNET and text encoder learning 1e-7. Image processing. Now this is really important. As higher as you go with training max resolution you will get better results, but max resolution also depends on your base trained model. Realistic vision has a very good resolution, but probably the max we can go is 1024 so make this 1024 like this. Do not select apply horizontal flip or dynamic image normalization. I tested this and they are decreasing the quality. Use EMA. Select Lion. I tested all of this and Lion is the best one that's for sure. Select mixed precision bf16. If your graphic card do not support this. You can also select fp16. I will continue with bf16. Do not use xFormers. This is also decreasing the training quality. Cache latents. Train UNET. Step ratio of text encoder is 75%. Offset noise is used to have more contrast of the output images. I don't find this useful for training yourself, but you can try this with like this. Freeze clip normalizations layers. This is not also improving so do not check it. Clip skip 1 for training yourself. Now this is important. Weight decay. This will be 10%. You see I did set a lower learning rate and higher weight decay. Why? Because I am using Lion optimizer. So these learning rates and these weight decays are set for Lion. So apply everything like I am doing. Do not touch anything else here. Keep them as they are. Now concept tabs. This is really important. In the dataset directory we will set our training images folder path. Here I have preprocessed different dimension images of myself. How did I prepare them? I used my human cropping script which I have shown in this video and prepared automatically different aspect ratios. For most realism we need the highest resolution with 768 by 1024. These are the raw images. They are bigger than 768 and 1024. Then I used the train tab. In here, preprocess images. Set the source directory. For example let's set this like this and here let's set another folder as a destination. So I will set the destination here. Let's make it as new train. Set the width and height of the images. Select auto focal point crop, focal point face weight 1, and preprocess. I am not using captions for training a person. It is not improving the realism. So here now we have our new training images. They are processed. They were already properly cropped for the correct aspect ratio. So when I click view details and sort by width and height, we can see the width and height of the images like this. They are ready to be used. So the most important part of DreamBooth training for realism is the training images data set. This data set is at best medium quality. Why? Because I have repeating backgrounds and repeating clothing. So as you have more diverse backgrounds and clothing, it is better. Whatever is repeating in the training images, the model will memorize them. Therefore you will have lesser flexibility. However, even this data set works very well for realism. So if you use a better data set then it will be better. In your images the sharpness of the image should be very good. It must have very high quality. It must have good lightning. So capture your training images during daytime when you have sufficient amount of lightning. This is really important. Your images should be sharp. Your images should be high quality. Sharp focus, clear, not blurry, not anything else. So the most important part, as I said, the training images quality. Prepare different backgrounds and clothing having images. This is at best a medium quality data set. Okay, let's return back to our DreamBooth tab. Let's select data set directory. So I copy here, paste it here and training images data set is ready. Then classification images. This is a commonly asked question to me. What is classification images for. The classification images will prevent model from overtraining and also it is used to keep the model previous knowledge. Therefore, normally you need to generate your classification images by using the model itself. Let's say we are going to use class prompt as a man and you may wonder, what is this man, what is this classification images and everything? I explain all these concepts in this amazing tutorial video. So watch this video to learn more about classification images and rare tokens and other concepts. Since our target is realism, I will fine tune the model with real images. So with the classification of real images, the model will become even more realistic and I will get better results. So here my preprocessed images. I will use 768 by 1024 pixel images. So here my data set. These are collected from unsplash.com. These are available to use even for commercial purposes. You see these are real images. They are super high quality. Let me open some of them for you. Here an example image. Here another example image. By using these real images, I will improve the realism of my model and I will prevent overtraining. So with these images, I will obtain the maximum realism from this training. This classification images data set is shared on this post. You can alternatively also watch this tutorial video and prepare your own classification images data set as well. Let's say you want to use the model itself classification images. Then you can generate classification images with photo of man, set the resolution the half of the target resolution so it will be 384 by 512. Then apply high resolution fix why? Because this way you will get higher resolution, better quality classification images data set. For example this is an example image but it has some noise as you are seeing with high resolution data set. So right click here and generate forever as many as classification images you want, then move them into another folder and give its path. However for realism, real images performing better. So I will copy the path of this classification images data set here. I don't use file words instance token or class token. I use training prompt so it will be ohwx man, this will be my instance prompt. Ohwx is our rare token. Man is our class token. So we teach ourselves into ohwx rare token and we leverage the knowledge of man token of the model. The class prompt will be also man. Because my classification images are all man. Let's say you also wanted to use captions for your images. Then all you need to do is adding here [filewords]. So it will read the captions and append them to here and this will be your final instance prompt when during training. You can prepare captions from train, preprocess and here you can use BLIP for caption or deepbooru for caption. For example, let me show you use BLIP for caption. There are also other captionings, but I find the captioning is necessary for fine tuning, not teaching a style, or for not teaching a subject. In this case, we are teaching ourselves as subject. Therefore, I don't find using captions is better. So new train2, preprocess. When you first time use captioning, it will download the used model. In this case it is BLIP and here our second training data set with captions. When you open caption file you will see a description, a man with glasses and a blue shirt is standing in front of a building and a street light and street light. So you see the class token is already exists in my caption. Therefore, if class token exists in all of the captions, remove it from here. So when this file is read, what will be the instance prompt. The instance prompt will be equal to this one. It will just append the file words to here. However, as I said, I don't find it better to use because it will diverse your training and it will take your model away from realism. In realism training, we want to model to learn as much as the details of our face. And the class prompt is man and class images per instance. Now this is important. I use 100. Therefore, since I have 13 training images, I need 1300 classification images. How many I have in my folder? I have 2700 classification images. Therefore, calculate your number of classification images that you need. Generate them from here. Open the folder, look how many you have, move them into another folder and use them as a classification images. However, you won't get as good as results as using the real images. Because when we use real images like these are. Actually we are further fine-tuning the base model with these new images for this token. So it becomes even more realistic. Better quality for man token. I don't use any sample prompts because I will check the each checkpoint with x/y/z plot I will show. So these are the settings of concept tab. Data set directory for training images. Classification dataset directory for classification, regularization images. Filewords these are empty. Training prompts ohwx man. Class prompt man. If you are training woman, then this is woman. If you are training girl, then this is girl. If you are training boy, then this is boy. Moreover, you should also clear the bad images from your classification images. In the saving tab do not save EMA weights for generated model. This is not improving the quality. Generate a ckpt file when saving during training. By the way, not anymore it is generating ckpt. It is generating safetensors file. It is same but this is the safe version. So these are the saving options. Always save settings, then click load settings. After loading settings verify that all of them are correct or not. Training steps per epochs, same model frequency: 10. Batch size 1. Gradient 1. Set gradients to none when zeroing. Gradient checkpointing not selected. Learning rate, text encoder rate, max resolution this is important. Make sure that it is 1024. Also, make sure that your classification images and training images are also at this resolution. If you don't increase this resolution, then the DreamBooth extension will downscale them first. Then start training. So you won't get this much high resolution quality. Use EMA. This is important. Lion bf16, memory attention default, cache latents, train UNET. Offset noise 0. Clip skip 1. Weight decay 10%. Text encoder weight decay 10%. I also tested this text encoder gradient clip normalization. This didn't improve my results. Concept tabs looking correct.Training images, classification images, dataset training prompts correct. Don't use other words here. Because rare token logic is very important. And as I said, if you don't know rare token logic, please watch this video. This will help you tremendously. Okay, everything is looking good then hit train. Currently my VRAM usage is like this as you are seeing right now. First it will start with preprocessing and looking for the classification images whether they are correct dimension or not and every one of them is in the correct dimension so it should work very well. First, it will look for the classification images and preprocess them if they are not in the correct dimension. It took like four to five seconds to preprocess my classification images. It says that it has found 1300 regularization images which was the necessary number. Why? Because we have 13 training images and we are using 100 classification images per instance image. Therefore, we need 1300 classification regularization images. Since it has found all of the necessary classification images, it didn't generate any new ones. However, if we didn't have sufficient amount of them, it would use these settings to generate those missing number of classification images based on the class prompt here. However, as I said, do not depend on this. Use text to image tab to generate your classification images. It is better. Always watch this cmd window to see what is happening. This rule applies both on windows on linux, on RunPod, wherever you are doing your training, always look at the cmd window that you started your web UI. Now it is caching latents. It is caching both classification images and training images. Therefore, 1313 images are being cached in the gpu VRAM memory. So caching latents will increase your VRAM memory usage. Caching latents has been completed and now it started training. You see the training will going to take about 6 hours to be completed for 200 epochs. It is using 19 gigabytes VRAM. Why? Because we are not using gradient checkpointing, we are not using xFormers and we are using 1024 pixels resolution and we are using EMA. All of these are increasing the VRAM usage and decreasing our training speed. However, these are necessary for maximum quality training, and when you consider that you don't have to go to a photo studio anymore, this is much more convenient both economically and from both timing and the results perspective. By the way, currently I am recording a video. Therefore it is also using a lot of GPU and VRAM of the GPU as well. Moreover, I am using Nvidia broadcast for reducing the noise and this is also using a lot of GPU. What if if you don't have such powerful GPU for such quality training, don't you worry. I have two excellent tutorials that shows you how to use Stable Diffusion and DreamBooth on RunPod. First one is this one and second one is this one. I even have automatic DreamBooth installer on RunPod. In this post it is shared. Follow this video to install DreamBooth on RunPod. For doing this training you don't even need a powerful GPU from RunPod. Just rent RTX3090. It is 29 cents per hour and with like spending two or three dollars you can have your amazing trained model, then download it and use it on your computer or on Google Colab. Wherever you want after you have done the training. Everything is same on RunPod. Just watch these two tutorials and you will be able to do everything I have shown in this video also on RunPod. Meanwhile our training is continuing. I had done exact same training previously me best. So I have a lot of checkpoints and I have no idea which one is the best. Therefore now I will show you checkpoint comparison. Select like the middle like 100 epoch like this. Decide your prompt. I will use this prompt. This is from the readme file. I will use this negative prompt. Now since we have done our training in 1024 you should select with 768 and height 1024. This is really important and that's all you need. Now generate some images and we will pick good seeds to do x/y/z comparison. So I am making batch count 100. We got the 100 output. Now we will select few good seeds and we will do x/y/z checkpoint comparison. So let's find some good seeds for comparing. Okay, this is a decent seed so I will look for the seed and save it. This is the seed. I copy paste it into a notepad file. Let's continue. Okay this is another decent seed. Let's also copy it. Paste it here. Continue. You can find up to 9 seeds or 4 seeds. It's up to you. Wow this is very interesting one. The face is almost like me but it has added more hair. I will test this. Here the seed. This usually don't happen when you aim for realism. The face is very good and it has also extra hair. Okay this is another decent seed. So I have got four seeds for experimentation. Then make the batch size 1. Go to scripts: x/y/z plot, copy, paste the seeds like this. Select the checkpoint names. From here you should select the checkpoints with the order that will be easier for you. So I will start from epoch 10. Why epoch 10 is 260? Because I have 13 training images. Therefore for one epoch it has to do 13 steps because in each step since batch size is 1, it will train one image. But I also have classification images. Thus, it is double of the required step. So 260 is 10 epoch, then 20 epoch, 30 epoch 40, 50, 60, 70. I am selecting all of the epochs with the order they are like this. Up to 200 epochs. So all epochs are selected. Make the grid margins like 100. It will be easier to see and nothing else. Hit generate. Now it will generate x/y/z plot comparison with the selected seeds so we will be able to see which checkpoint is performing best. You see in the first checkpoint. It is not even like me. As the training continues and more epochs are trained, the images will become more like me. When doing x/y/z comparison, make sure that the VAE file and the model files are correctly loaded. There aren't any errors. Also, if you have noticed I am using VAE ft 840000 ema pruned ckpt as VAE file. This is the best VAE file. It will significantly improve your image generation output. This is not used during training, it is used during inference. If you don't know how to install and use that VAE file. If you don't know how to set these quick settings, then watch the first 15 minutes of this tutorial and you will learn all of them. The x/y/z plot generation is completed. When we click the image, it will show us the entire image. However, looking it from the folder is better. Click this folder icon. You see it has opened text to image images. But we need to check out text to image grids. Here. The grid file will be saved here. So in the very bottom we see our latest generated grid file. Let's open it inpaint.net and let's zoom it to original resolution. So it starts from the epoch 10. Let's look epoch 10 nothing like me. This is epoch 20. This is epoch 30. In epoch 30 a little bit like me. Epoch 40 you are seeing right now. Epoch 50. Okay, epoch 60. The face is very similar to me. 70 yeah, not bad. 80. This is epoch 80. Epoch 90 as you are seeing right now: epoch 100. This is the epoch 100. Epoch 110. This is the epoch we used to find these seeds as you are seeing right now they are the same, okay 120 epoch. So by looking all of this you need to find your best checkpoint. Then we will use that checkpoint to generate images. You see as the epochs goes more it is changing and it is more becoming like me as you are seeing right now. This is epoch 150. Actually, this was the epoch I found best. Then 160 as you are seeing right now. So the quality begins degrading I think. You see the face is better here than here I think. Also better here than here. So you need to find your own best epoch. This totally depends on your training images data set, therefore I can't say it will best with you. However, up to 200 epochs is usually sufficient to find a sweet spot. Actually, after 110 I think all epochs can be used, but you see now we are starting to get some noise at the epoch 200 so it started to overtrain at the 200 epoch. That is for sure. That hair is also gone, the face is degraded, so that is why I prefer to go to up to 200 epochs and then find the best one with this way. I think this is the best epoch 150. The face is there. It is very high quality. After you decide your epoch, select it from here. Turn off the x/y/z plot, then hit generate forever. You can also add different prompts here. Try different prompts but why generate forever? I will explain you the logic of that. So in this folder I have over 2000 generated images but they have a different numbering and there are some values you are seeing there. These images are sorted based on the similarity to my real images. This makes my job much easier to find higher quality images. Much easier. So by sorting images based on the similarity and by generating hundreds or thousands of images, you can get very high quality images very easily that you can use in your Linkedin, Twitter, or wherever you want. Very like you with very high quality. Like they are taken in a studio shot like they are a professional photograph by wearing high quality suit. So how did I sort them. In this tutorial video I explained how to sort all of the images. Everything is explained here. I already have a Patreon post that includes script but you don't have to be my Patreon supporter to obtain the scripts. Everything is explained in this tutorial very clearly. So can these images be further improved? Yes. You have inpainting option. There are two ways to improve generated images. However, when you use inpainting I think it is not reaching the level of the raw output. Therefore, hopefully I am planning a fine-tuned model with better resolution than the realistic vision and do higher resolution training on that new fine-tuned model. Hopefully I will make a tutorial video for that and release the model as well. So with that model, hopefully we will be able to do training with even higher resolution and obtain better quality. But for now with realistic vision this is the resolution I think we can go up to. Click send to inpaint. There are two options of inpainting. The first one is using the native inpainting of the Automatic1111 web UI. Select the face select only masked and very important thing here is only masked padding pixels. This will determine your output. This will affect your output significantly. So let's try with default values first. Okay, this is the default values first try as you are seeing. Let's open it in a new tab. Then let's make the padding pixels 64. As you increase padding pixels, you will likely to get better image. However, there is a sweet spot so this is 64 pixels. You see from this to this. This is looking better. Then let's try 96. So you should try different padding pixels until you get the best result. Okay here it is. From this to this to this this is 96. This is 64. 64 is so far best. Let's say 128. By the way these are not cherry-picked this is the first result that we are getting. So with doing multiple times generation you can get even better ones. This. So this is 128. This is even looking better. Okay. But you see you don't get that natural look when you do inpainting. I think most natural look comes from raw output but this is still decent. So this is one way to improve. What is the other way of improving. By the way, this is a real improvement. The second way of improvement is using ControlNet. I have shown how to do this in another tutorial but I will also show here. First click enable, click pixel perfect, select tile from here. This doesn't produce always better results but you can also try this and see if it will produce better for you. Select tile color fix, select my prompt is more important. This is the first ControlNet then select ControlNet unit one. Enable pixel perfect. Select inpaint from here and don't change anything. This is important. Select whole picture and select inpaint not masked okay and hit generate. This will inpaint the entire image and it will likely to improve your face and also clothing. However, the best output is always from raw images. This doesn't work. Let's also increase the denoising strength to 1 and try again. Okay, this is a little bit better, but still not good as the original inpainting. Sometimes it produces better results, sometimes it is not producing better results. So I have an example here. The left one is the original, the right one is the raw output and here we are seeing the ControlNet applied option. I think it is subjective which one is looking better. So it is up to you. I have done a lot of testing and sometimes it is producing better, sometimes not so it is up to you to test. I think with this way it is adding more details like you are seeing here from here. But still both of them are good. So it is up to you to also apply this ControlNet trick. So what is important here is as much as possibly getting a good base image. Good raw output then inpainting face to fix. This is also very high quality inpainting result. Let me show you the settings one more time. Inpaint masked, masked content original, only masked, only masked padding pixels 96, resize is same, didn't change the resolution. Denoising strength is 75 percent. This is all the settings. So this is the inpainted result. Let's compare them. This was the base image. So from this base image to this inpainted result which is a very significant improvement. It is also looking pretty natural. Let's also compare it with the real image. So here on the left we have the real image. On the right we have the inpainted face. However, we are able to get this quality because we are doing training on a realistic vision model the most realistic model available right now. We are using real class images and we are using very high resolution. All these are affecting our results significantly. So I will put the summary and everything explained in this video to this Patreon post. You will be able to find everything on there. Alternatively, you can watch the entire video again. I am still going to work on even a better workflow for more realism. I am going to work on styling workflow as well because with realism and styling you can't get both of them at the same time. You can get either good styling or you can get either good realism. Moreover, I will work on a new, better, fine-tuned base model to train our images onto them to obtain even better quality images. Better output. If you get any question, just reply to this video. This is all for today. I hope you have enjoyed. Please subscribe, support me by joining on Youtube. I would appreciate that very much. Please consider supporting me on my Patreon page. This is super important for me. We have over 250 supporters. I appreciate them very much. Supporting me on Patreon will also help you tremendously. It will make your life easier. You will be able to access the scripts. I explain everything in my video so it is not mandatory to support me on Patreon, but if you support me on Patreon I would appreciate that very much. So as I said, this readme file will get updated as it be necessary. Everything you need is written on this readme file. Please read this readme file very carefully. Follow the steps, join the Discord channel and ask any questions that you might get. Hopefully see you in another amazing video. By the way, I have written the used commit hashes in the very bottom of the readme file. Moreover, I am sharing pip freeze command. With this command you are seeing all of the libraries and their versions. If something gets broken, you can install specific version of specific library and fix the error. However, this is probably won't be necessary. Just reply to the Youtube or the Patreon post and hopefully I will update this readme file or the Patreon post.
Info
Channel: SECourses
Views: 15,994
Rating: undefined out of 5
Keywords: dreambooth, dream booth, stable diffusion, stablediffusion, training, auto1111, auto 1111, automatic, automatic1111, automatic 1111, web ui, sd web ui, sd, lora, model, tutorial, guide, dreambooth training, dream booth training, lora training, realism training, person training, dreambooth extension, dream booth extension, studio, photography, ai, ml, artificial intelligence, machine learning
Id: g0wXIcRhkJk
Channel Id: undefined
Length: 42min 21sec (2541 seconds)
Published: Sun Jul 02 2023
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.