
Getting Started with DeepSeek: The Best Options for Every Use Case In this video, I give a quick overview of the best places to get started with DeepSeek. Whether you're aiming to conduct research, use it within a coding context, or run it locally, I've got you covered! 📝💻 Here are some key highlights: 1️⃣ Hosted Interface: chat.deepseek.com - Easy access but currently facing high demand. 2️⃣ GitHub Marketplace: Explore different DeepSeek models, including R1 and O1, on Azure. 3️⃣ Local Models: Check out Olama and LM Studio for offline use. 4️⃣ Jan: Provides a sleek chat interface and local server capabilities. 5️⃣ Perplexity: Great for research with seamless reasoning and citation features. 6️⃣ Groq: Fast inference with the 70B parameter model. 7️⃣ Artificial Analysis: Compare models based on quality, speed, and price. 8️⃣ IDE Integration: Tools like Continue, VS Code, and others make coding with DeepSeek easier than ever. Don't forget to like, comment, share, and subscribe for more awesome content! 🚀 00:00 Introduction to Deep Seek Options 00:14 Exploring DeepSeek Hosted Interface 01:00 Using DeepSeek on GitHub and Azure 01:37 Running DeepSeek Locally with Olama 03:04 LM Studio: A Local Interface Option 03:36 Jan: Combining Local and Server Capabilities 04:12 Perplexity: Research and Reasoning 05:30 Groq: Fast and Powerful Model 06:29 Artificial Analysis: Comparing Models 08:17 Deep Seek in Coding IDEs 09:48 Conclusion and Final Thoughts
Weekly deep dives on AI agents, coding tools, and building with LLMs - delivered to your inbox.
Free forever. No spam.
Subscribe FreeNew tutorials, open-source projects, and deep dives on coding agents - delivered weekly.
--- type: transcript date: 2025-01-30 youtube_id: n9Ir-nDZxy4 --- # Transcript: Get Started with DeepSeek in 10 Minutes in this video I'm going to do a quick overview on the different places on where you can get started with deep seek whether you want to do research if you want to use it within a coding context if you want to just use it locally I'll show you all of the different options that are available right now the first one and this is probably the one that you might have already visited and this is chad. deep seek.com this is the hosted interface where you can interact with the model so you do have to enable deep think R1 to access the R1 model and what's interesting with this model is you can ask it about recent events so you can say tell me about today's AI news and what it will do is it will actually perform that search functionality for you before feeding it to the model now mind you one thing that's interesting right now is they seem to be having troubles with actually the inference and serving up the amount of Demand right now I think that is in part why we've started to see a bit of a decrease especially over the past day not being able to handle the amount of requests that they so GitHub has a Marketplace if you go to g.com /m Marketplace you'll be able to see the different models here so there's deep SEC car1 there's 01 you can go in and you can try this and for instance just if I show you this now this is running on azure's infrastructure and when you compare this to the Deep seek hosted model if you have been able to try it you will notice this model is very slow in comparison that's one thing to note if you are going to be using this through GitHub or on Azure it does seem at least at time of recording it is going to be a little bit of a slower option now the next one is going to be for local models so olama is a really great option you can go to ama.com you can pull this down and to get started for the models you can go to the models page here and on the specific deep seek R1 page now there's a number of different sizes of models if you're less familiar with language models they're going to come in various sizes often now what's different with this release is the main R1 model with all of the performance that people are toting in comparing to the 01 model from open AI this is that 671 billion parameter model this model is not going to run on standard Hardware you're going to have to have Nvidia gpus or you're definitely going to have to have some specialized Hardware to be able to run that so now in terms of the most common sizes of what you'll be able to likely run on your machine are generally speaking I probably recommend you start somewhere with the set 8B variants of the model you could go as small as 1.5b generally speaking I think most people should be okay with 7 or 8B now if you do have a decent computer you should be able to run 14b now the one thing that will vary machine to machine is the tokens per second rate will definitely fluctuate a fair bit depending on your machine now if you have a really good machine with a lot of ram you should be able to go up a little bit on the Spectrum here as well if if you want to try some of these more powerful models next up another great option is LM Studio this is a great one where you can download it on Mac windows or Linux similar to AMA but what's different with this is you actually have an interface as soon as you download it you'll be able to pull all of the different models that you want from hugging face that are supported and you'll be able to have this nice little interface and be able to run it all locally that's the big benefit of both AMA and LM studio is you could just turn off Wi-Fi on your computer and no information is going to be sent across the network you can use these completely offline so next up is Jan and I really feel like this one sort of sits in between the capabilities of AMA and LM Studio the great thing with this is you have a really nice chat interface but they've also added the ability where you can use this and serve it up as a server that you can access say if you have local applications and you want to integrate you can just make requests to your local server to be able to use those where that can be useful is if you're using a coding IDE and you want to make requests to a language model that you have locally you can do that with things like Jan LM Studio as well as Olana next I want to show you a really great implementation within the application layer that came out very quickly perplexity rolled out in deep seek R1 seemingly in record time and what you're able to do with this is if you ask a query in this let's just say the news today was open AI claims de deep seek used its model what this will do is you will see it will go through the research step and what's great with perplexity is it's actually built for research by being able to incorporate a quote unquote reasoning model it can reason about all of the different information that is searching for and it can go and think through it so you'll see okay let's tackle this query the user has asked about opening eyes claim that deep seek use his models first I need to go through all the provided search results to gather key points so on and so forth then with in here we can see these sources and then it gives us this aggregate response with all the citations in line and everything kudos to the team perplexity for integrating this it is definitely a really slick look and being able to use these reasoning models that can access the internet in my opinion is like the killer use case right now with this model there's definitely a gentic use cases that I think we're going to see over the coming weeks but right now the thing that works really well is that search and reasoning capability combin so next up is Gro so grock is serving up the 70 billion parameter model which is the second biggest model you're going to be able to use the model incredibly fast if I just say write a short story and I'll submit this this is hosted and served up and what they use to make it so fast is right now at time of recording the inference speed from grock is about 275 tokens per second and now this is on the 70 billion parameter model this is a distilled version of of that model which is still very powerful just to demonstrate it here I put out a really quick demo when they put this out and to show you how this can work is this is effectively a clone or a light clone of anthropics artifacts feature and as you can see here it is very quick to render everything it just ran through the answer it followed all of the system instructions that I had in place and it was able to give this coherent response without any syntax breaking or anything now the next thing within this next where I want to point you and this is definitely a good one to keep your eye on is artificial analysis within artificial analysis what's great with this is they rank all of the different models across quality speed as well as price you can see all of the different hosting providers as well as the rate of things like the latency their particular context window and all of these different metrics such as cost which do vary considerably platform to platform and what's interesting with this is actually the cost in comparison with this speed across the board for the Deep seek hosted version now mind you this has really been hammered the past several days it is going to be interesting to see if they're able to keep the prices this low because it might be a little too much to actually be able to serve up this model especially with the amount of attention that it has right now now in terms of the output speeds for the Deep seek interface it is actually very fast it is an interesting question considering what they might actually be doing not just on the training front but also on the inference front to be able to serve this sub on gpus at a speed like this because when you compare it to some of these other providers like hyperbolic CML fireworks together AI these are all the infrastructure companies that run with Nvidia gpus and for the second best Contender to have performance that's less than half of what deep seek has achieved it's an interesting question there on how deep seek was able to get the inference speed this fast there are a ton of other metrics here that you can check out they do track the out put speed over time when a model comes out you of to see that the API speeds do degrade a little bit especially after a major release and this can just be helpful overall for helping you determine what might be a good provider to use or potentially if you're going to be using multiple providers next I just wanted to point you to a couple other options you can set this up in a number of different idees right now one that I personally quite like is continu you can use this in vs code and what you'll be able to do with it is essentially have a cursor like experience where you have this panel on the right hand side and you'll be able to make edits and suggestions to your code with natural language what's great with this actually is you can integrate it with olama for instance I believe you should be able to integrate it quite easily with LM Studio as well as Jan if you'd like and what you can do with this is not only just leverage local models but you could also use a provider like Gro if you wanted or if you wanted to use any of those providers like a showed you within artificial analysis now in terms of some other idees I've had recommendations to check out Klein this is one that I've heard more and more about I haven't checked it out myself so I can't say too much on it just quite yet but this is an increasingly popular option that people like for being able to use this across a number of different models within VSS code now another option that you can try this out in is ader personally I haven't tried this quite yet but I know a lot of people do quite enjoy this experience you'll be able to use this within your terminal and be able to build out your application with natural language they also have some really great benchmarks for coding on how well these different models perform when they come out I've definitely touched on those before but this is another potential option that you could use if you're looking to use it within a coding context otherwise that's pretty much it for this video I just wanted to show you a number of different options especially if you wanted to try deep seek but maybe you haven't been able to try out the Deep seek interface quite yet because it is just being hammered by the looks of it otherwise that's pretty much it for this video if you found this video useful please like comment share and subscribe otherwise until the next one
Technical content at the intersection of AI and development. Building with AI agents, Claude Code, and modern dev tools - then showing you exactly how it works.