Run Llama 2 Locally (Quick & Complete Tutorial - No GPU needed)

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
Okay, straight to the point, let's begin with  downloading the text-generation-webui tool you can find the links below, as I'm using a Windows  operating system I'll choose the windows version after downloading, we'll extract the files and open the "start_windows" file as administrator to save time I'll speed up the video a bit,   this process might take anywhere from 5 to 20  minutes based on your hardware and network speed Here, I'll press D to run it in CPU mode next, we need to download the  model, you can find the links below I'll go with the 13B model model if you want   to use it in CPU mode you have to  choose the one that ends with GGML you'll find files available  here along with their use   cases you can choose the one that suits your needs Okay, let's click files and  versions and download it,   I already downloaded, so I'll navigate  to the "text-generation-webui" folder then the "models" folder and paste the downloaded model there okay, let's access it through  our local URL on the browser,   in the model section we may need  to refresh it to see the model (don't forget to load it) If desired, we can adjust parameters like   the temperature to control the  randomness of the generated text let's try it yep, it's working you can change the mode here, in my  observation, it doesn't surpass GPT-3.5;   however, if you require a local LLM model, it  could be a viable option. That's all there is   to it. If you have any questions, feel free to ask  in the comments section. Thanks for watching!
Info
Channel: Rhineverse
Views: 3,709
Rating: undefined out of 5
Keywords:
Id: Uq4TLnEPoI4
Channel Id: undefined
Length: 3min 53sec (233 seconds)
Published: Fri Jul 21 2023
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.