I've installed the new deep seek R1 reasoning model the large language model that Rivals open ai's 01 model which right now I'm paying $200 a month for this is completely local running on my computer and it's private I could literally turn off the Wi-Fi right now so in this video I want to show you exactly how to install it step by step and I'll make it as non-technical as possible now I'm also making a video comparing deep seek R1 versus open ai1 which I'll release in a couple of days but in the meantime I'm going to show you exactly how to install it so if you go to deep seek. com you could actually use their chatbot so if you click Start right now it brings you to this page and this is just available at chat. deepsec and you could create an account and then you could use it here that's not what I want to show you here I want to show you how to install it locally on your own device so you're not using a website and if you click deep think this is the R1 reasoning model now if you just have a normal chat with it it uses a different model not R1 it's called V3 deep seek V3 and that's been available since December and he also has the power of search so you could combine this deep think R1 model with search here too again I'll cover this in a more Deep dive video about the model right now let's go ahead and install it locally okay it's only going to take you a couple of steps to install it on your computer and then it's going to take a couple more steps to get a really nice user interface so it looks like chat GPT so the very first step of the process is we need to install something called olama on our computer at ama.
com this works for mac and for PC and just press download right over here choose your operating system from these three and then press download over here if you're using Windows it needs to be Windows 10 or later okay once you unzip and install AMA it's going to give you this popup right here I'm going to press next it's going to say install this command line and then you can press install right over here I'm going to type in a passcode then it's going to say run your first model it's going to give you a little code here just three words to use as your first model right now llama 3. 3 is out so I could actually change that to 3. 3 so I'm going to copy this and I'm going to press finish right here now let me before I installed R1 I'm just going to open up and search for the terminal app over here now this is optional but I'm going to install llama 3.
3 so I'm going to type in 3. 3 here so it says AMA space run space llama 3. 3 I'm going to press enter this is going to install llama 3.
3 which is obviously not R1 but this way we have another open source model that is not a reasoning model so we could go back and forth in the rest of this process while this is installing let me show you R1 by the way llama 3. 3 is 42 gabt in size so make sure you have that kind of space here on your computer if you're going to do this if you don't have that much space skip this and just installed R1 which we're going to install pretty much exactly the same way now we're on step two now we have Ama installed we're going to go back to the ama. com website we're going to click on models and we're going to see this deep seek R1 model available right here let me click on it so R1 comes in multiple different weights and sizes okay so if I click this you'll see that it comes the smallest version is 1.
5 billion parameter now that's not going to be very good but it's going to be very fast and very small it's only 1 Gigabyte by default is set to the 7B model and then you could go all the way up to 671 billion parameter model which is 400 GB and no computer is going to be able to handle that at least no consumer grade computer and depending on how good your computer is you may be able to get to that 70b model but I usually recommend for these type of things start with the 7B model and then see how it is with speed because it's going to require pretty beefy GPU your graphics card in order to run this properly now I also got this new GPU this is called GeForce RTX 590 top of thee line GPU available right now from Nvidia this is not sponsored by them but they flew me out to CF and I made a different video about the different things they rolled out there but I'm going to try to build a new pc and I'm going to install this in there so I could take advantage of some of these other models so I could show you the local install of those so I'm going to start with the 7B model myself so in order to install that and we're just on step two installing the large language model using AMA I'm going to copy this right over here so it says AMA space run space deep seek D R1 that's the code I need and if you you were to install the other models all you have to do is click on it here and then you see it just adds that to the end of the code so it adds 14b and then you can install multiple models too and then when we get to the user interface that we're going to use for this we're going to be able to choose between different models too but right now just get started with one and I'll copy the code from here which is by default the 7B parameter model okay so I restarted terminal so I could start here fresh I did install llama 3. 3 before but I'm going to paste AMA space run space deep seek D R1 again which is what I got from that ol website and I'm going to press enter here and now it's going to go ahead and install that this is 4. 7 gabt this version of it which is the seven b version the smaller version here okay and that took about a minute here that depends on the speed of your computer obviously because it's downloading it and right here I technically right now have deep seek R1 running on my computer this is how I could use it inside of the terminal app and I could literally ask it anything it does its thinking right here how can I assist you okay so I'll ask it how many RS in strawberry and this is how it works it creates that thinking bracket if you've used chat gpt1 Works in a similar way now it actually didn't give me a complete answer so let me try again and this time there are three letter RS in the W strawberry and it kind of broke down its thinking process here now obviously this is all you need to to do to install it locally but it's not very nice right we need couple more steps to install a user interface that looks a lot more like chat GPT or the Deep seek chat bot I showed you in the beginning but the first thing we need is we need to install something called Docker this is a doer.
and then again we could download this for our computer and I'll move it to my application folder now I just need to search for it on my computer after it's installed and I just need to double click to open it and it just needs to be open I literally don't have to do anything else and I'm not even going to create an account so I'm going to skip this okay and then you should see this page where it says your running container show up here and there shouldn't be a container just yet and we don't have to do anything with Docker it just needs to be installed opened and minimized so I'm going to minimize it now now I only have one more step so the first thing just to recap real quick so we're on the same page we went to the AMA website we downloaded AMA we went to the models tab we went to deeps R1 we copied this over and installed it inside of the terminal app that's all we did and then we went to the docker website and then we downloaded Docker from here depending on what machine we have and we installed and minimized it now the last thing we need to do is we need to install something called open web UI and this is what it's going to look like it's going to run llama 3. 3 it's going to run deep SEC car 1 any model and we could pick and choose what model we want depending on what we're doing with it and I'm going to show you exactly how to install it now this page docs openweb ui. com I'll include this in the description of this video with all the steps so all the links are going to be in one place and all I have to do right here it says if AMA is on your computer use this command and it gives you this Command right over here so that's all we have to copy from right over here I'm going to copy this code and let's open up terminal one more time I'm going to actually terminate this terminal and start a new one okay and then I'll go ahead and paste that code over here and press enter and it looks like nothing happened but it did create something inside a Docker now so let me open up Docker which I had minimized over here and you could see now I have something called open web UI and this is a container and it gives me this port number right here which is going to open up a tab inside of a browser but it's offline this is actually not going to be online right here and then from this page right here we do need to sign up for open web UI to get that UI and then after that your computer could be offline it doesn't have to be connected to the internet to use it I already have an account so I'm going to sign in okay now that brings us to this page right here and this is what I'm want to show you now where we could use R1 so the way we do that is right up here if you click this dropdown you'll see deep seek R1 latest the 70b model is right over here so I have different llama models that I have installed I have older models here I've had this for a while but if I click on this right now this is going to now use R1 if I set it as default every single time I use it it's going to use R1 but most of the time I actually probably want to use something else like llama my 3.
3 which is what I installed in the beginning of this video maybe that could be my default because it's more Broad and not a reasoning model and then use R1 when I need it but in this case since it's a deep SEC car1 video let's go ahead and use it right over here okay let me give it this prompt right here give me 10 challenges only a reasoning AI model can solve and I'm going to press send I'm not going to edit it so you see the speed of it right here you see that it's almost instant because it's a smaller 7B model that I'm using the higher the model the better but the slower depending on your computer here and you could see it always puts it in these thinking brackets so it starts a bracket and then it ends a bracket so this part of it is actually not your answer this is how he thinking through it and then down here this is going to be your answer so he gave me 10 different things problem solving ethical decision making strategy games mathematical problem solving so this just gives you an idea of the different things that you could actually do and one of my favorite reasons for using open web UI is this right here right on top I could choose this model right here but I could actually select the secondary model so I could press the plus sign select another model since I installed llama 3. 3 I could run those side by side ask it for the same exact prompt again okay and look what it does now he actually uses both models side by side you could see llama 3. 3 is like 10 times the size right now that's the much bigger that was 43 gigabytes this was four and you could see how much slower that is and this is why I probably need a much better computer here to run these models locally as they get bigger because I mean it's been what like 10 seconds and I'm still not getting a response out of llama but R1 went to work right away okay now it's finally going to work it's very slow typing out kind of one word at a time but I'm running this on a laptop right now so the fact that it is even running is kind of surprising here this llama 3.
3 but you can see how fast R1 is and then I could install more models so let's say I wanted to install the bigger model since 70b worked well let's try 32b at this point is super simple I just go back to the models tab on the AMA website click on 32b over here and click on the dropdown copy this over open Terminal paste this into terminal right here press enter and then it's going to go to work this is going to be 19 gigabyt so it's going to take much longer than the first version I installed here and then as soon as it's done let me show you what ends up happening inside of this open web UI and if you look at my URL by the way I'm on the browser I'm on Chrome right here but I'm not on the web I'm on Local Host 3000 this is running locally right now on my computer now while that's running I want to also show you this because we recently revamped our entire e-learning platform for AI and we actually merged with futurepedia which is one of the leading AI tool libraries and newsletters and now it's a kind of a all-in-one type of platform and we have a course specifically for this right here AI powered private chatbot and this one is about an hour but it really dives into all the different things you could do with running llama models running mystal models learning different models now I'm going to add a little section here about running R1 models but it's focused also a lot more about the open web UI interface about actually giving your knowledge base to it all kinds of different things you could do inside of that so that's available here and this is all in one subscription so not only do you get this you get all these different courses notebook LM course comes out tomorrow so all these are going to be available under the same bundle and we still have a free trial available right now so if you want you could actually watch a course for free see if you like this platform there's a whole Community section all kinds of different things here you could explore once you become a member just to see if it's a good fit for you and then you could complete courses and get certification and things like that all in this platform I'll link this below as well okay this one just finished up installing so I'm not going to use it inside of terminal I'm actually going to close terminal here and now if I click this drop down it's not going to appear just yet over here but if I refresh this tab over here now if I click down here I'm going to get two different R1 models so now we have a 32b model and we have that 7B model so I'm going to click on this one instead this time and let me actually start a new chat okay let me choose the other one here and in this case I'm going to run the other one I just want to show you kind of the speed of it here so this is 7B versus 32b I'll paste the same prompt it's is a relatively simple prompt here right it's not really doing any reasoning okay that was not bad it actually kind of worked wow look at this this is really surprising that the 32p model is actually working well not exactly as fast but I mean it's not like the Llama 3.