With the rapid advancement in Large Language Models (LLMs), Deepseek-r1 has emerged as a top contender. Matching the performance of the GPT-o1 model in reasoning and code generation, Deepseek-r1 leverages transformers and logical tree reasoning to optimize decision-making, setting a new benchmark in AI-powered development tools.
Having used tools like Cursor and other paid code helpers, I was curious to see how Deepseek-r1 would perform in my workflow. This sparked my journey to integrate it into my environment and evaluate its capabilities.
Let’s Get Started 🚀
Step 1: Install Ollama
To get started, you’ll need Ollama, a platform that allows you to run LLMs locally. If you don’t already have it installed, download it from Ollama's official site. Follow the setup instructions to get it running smoothly on your machine.
Step 2: Download Deepseek-r1
Once Ollama is up and running, it’s time to pull the Deepseek-r1 model to your local environment. Run the following command in your terminal:
$ ollama pull deepseek-r1
After downloading, you can test if the model is working correctly by running a simple curl command:
$ curl http://localhost:11434/api/generate -d '{
"model": "deepseek-r1:latest",
"prompt": "Why is the sky blue?"
}'
If the model is successfully set up, you’ll see a chunked output in your terminal—indicating that Deepseek-r1 is ready for action.
Step 3: Install the Continue.dev Extension
With Visual Studio Code as a prerequisite, head to the Extensions Marketplace and install the Continue.dev extension. This extension bridges the gap between VS Code and advanced AI models, enabling seamless integration for coding assistance.
Once installed, you're one step closer to unleashing the power of Deepseek-r1!
Step 4: Configure Deepseek-r1 in Continue.dev
Open Continue.dev by clicking on its icon in the Activity Bar of VS Code.
In the chat window, locate the model selection button at the bottom-left corner.
Click on the button to open a dropdown menu.
From the menu, select Ollama as the platform.
Once Ollama is selected, you’ll see all available models in your local environment. Select Deepseek-r1 from the list.
At this point, Continue.dev is configured to use Deepseek-r1 via Ollama.
Ready to Code!
With everything set up, you can now take full advantage of Deepseek-r1's advanced reasoning and code-generation capabilities directly within VS Code. Continue.dev enables additional context-aware features, enhancing the development experience:
Autocomplete: Get precise suggestions for completing lines or blocks of code.
Code Refactoring: Highlight sections of code and ask for optimizations or rewrites.
Code Explanations: Select any code snippet and let the AI break it down for you.
Here’s a quick look at what the workflow will look like:
Why Deepseek-r1?
Reasoning Excellence: Logical tree reasoning ensures smarter, more optimized decisions.
Transformer Power: Delivers robust performance in code generation tasks.
Local Execution: Run models locally for better privacy and faster responses.
By integrating Deepseek-r1 with Continue.dev, you unlock a powerful development workflow tailored to your specific needs.
Cheers to smarter coding! 🥂
Top comments (59)
DeepSeek cut-off date is end of 2023, which is ancient compared to Gemini 2.0 (August 2024).
It is imperative to have a recent cut-off date for front-end related work or when querring anything about recent things.
This is why I will NOT use it.
On the other hand, Gemini is rubbish, so there's that.
Agreed, based on personal experience.
Rubbish?? its on the very TOP of the leaderboard:
lmarena.ai/?leaderboard
It could be on top of every leaderboard in existence, it's still inferior for coding tasks than say Claude (and since this is dev.to, that's the relevant bit).
It's on the top for coding in the world's most famous AI leaderboard website. it has been ranked so by at least thousands of developers, so this contradicts your one and single personal experience completely.
Whatever you say buddy.
Gemini is BERT which is bidirectional model not best for Q&A usage like other decoder only model, e.g. GPT , llama , etc
So how do you explain thousands of people had rated it so highly (for coding)?
I don't think they are all Google workers who are trying to boost their own product by rating it.
This will not run Deepseek r1, but rather Deepseek coder. Even if you wanted to run Deepseek r1, you would need a lot of processing power — r1 can't just run on a laptop or PC.
r1
comes with a lot of variants:The
1.5b
variant could very easily run on a weak or older device, and the8b
variant works fine on my device.I am running deepseek-r1 on my old intel macbook
Which variant? I doubt the full 671b variant works on a macbook. And everything under 32b is really only useful with fine tuning for specialized tasks.
Yes
for windows follow ;
: Setting Up Deepseek-r1 (Using Windows CMD)
Install Ollama
Download the Deepseek-r1 Model
To test if the model is working, use this command:
If you see output in the terminal, Deepseek-r1 is ready to go!
** Setting Up the Continue.dev Extension**
Install Visual Studio Code (VS Code)
Install the Continue.dev Extension
Connect Deepseek-r1 to Continue.dev
: Using Deepseek-r1
Now you can use Deepseek-r1’s features directly in VS Code:
: Why Choose Deepseek-r1?
: WhatsApp Status Message
"Deepseek-r1: Smarter than GPT, excelling in reasoning and code generation. 🚀
DeepseekR1 #AI #Coding #WindowsCMD"
Start using Deepseek-r1 in your workflow and enjoy smarter, faster coding! 🥂
(hashnode.com/@askpravinjadhav)
Most obvious question: why would you use this instead of CoPilot? or am I comparing apples and oranges ;-)
Maybe to not send data to a company you can or cannot trust
"A company you cannot trust" - DeepSeek is Chinese, does your "data" go straight to the Communist Party? ;-)
Not if you run their models locally...
There was a better guide a few days ago already:
dev.to/shayy/run-deepseek-locally-...
What I don't understand is that how come a small startup is able to build LLM model that beats chatgpt? Even Tech giants like Google or Elon Musk didn't build LLM model that can beat chatgpt. All of a sudden a small Chinese company comes up with LLM model superior than OpenAI's chatgpt. Did they use a LLM recipe which OpenAI doesnot know?Guys if there is something i missed here, please enlighten me😊
This is what you get when enthusiastic smart people try a different angle instead of just running for investors money. Yes, deepseek R1 is trained differently. They used a rule/heuristics based reward system and automated benchmarking during the training process... nothing new, but apparently nobody else thought of combining this with LLM training.
Hey Varnit, had a question for you after seeing this. How good is deepseek in terms of hallucinations, as yesterday night I was reading on LLM hallucination detection.
Don't you have to have a pc of decent memory & processing capibilities to do this?
not really there a variants with parameters they are trained on below 16B params most of the domestic use machines can handle
Huh... I did not know that.
gff12
gffg