随笔 - 27  文章 - 16  评论 - 1  阅读 - 36772

deepseek-r1部署

1.首先安装ollama (https://github.com/ollama/ollama)
直接按官网说的下载特别慢:curl -L https://ollama.com/download/ollama-linux-amd64.tgz -o ollama-linux-amd64.tgz
可以先加速下载下来,然后手动安装:sudo tar -C /usr -xzf ollama-linux-amd64.tgz
启动:ollama serve

2.在ollama官网搜索下载模型,比如:ollama pull deepseek-r1:70b
模型自动下载到~/.ollama/models/blobs/
模型在shell运行:ollama run deepseek-r1:70b


以下用法参考https://zhuanlan.zhihu.com/p/7736407412
对于ollama的api写法各有不同,其实就使用而言,最简单的request方法就可以奏效(注意在/api/generate 下,是没有“message”字段的)。

import requests
url = 'http://localhost:11434/api/generate'
def get_request(query):
data = {
"model": qwen2.5:7b,
"prompt": query,
"stream": False

}
response = requests.post(url, json=data)
return response
但是这里没有关于生成参数的设置方式,比如,某些场景下,我们不希望模型生成随机答案。又一番百度之后,发现其实传入参数还有一个options可以设置:

gen_option = {"temperature":0.0} #设置生成参数,取消随机性
data = {"model": "qwen2.5:7b",
"prompt": query,
"options": gen_option
"stream": False

}
response = requests.post(url, json=data)
return response
参考https://gitee.com/mirrors/ollama/blob/main/docs/modelfile.md

3.安装pyenv和python11
用https://github.com/pyenv/pyenv来管理python版本
先安装必要组件:
sudo apt update; sudo apt install build-essential libssl-dev zlib1g-dev \
libbz2-dev libreadline-dev libsqlite3-dev curl git \
libncursesw5-dev xz-utils tk-dev libxml2-dev libxmlsec1-dev libffi-dev liblzma-dev
然后
curl -fsSL https://pyenv.run | bash

add the commands to ~/.bashrc by running the following in your terminal:

echo 'export PYENV_ROOT="$HOME/.pyenv"' >> ~/.bashrc
echo '[[ -d $PYENV_ROOT/bin ]] && export PATH="$PYENV_ROOT/bin:$PATH"' >> ~/.bashrc
echo 'eval "$(pyenv init - bash)"' >> ~/.bashrc

Then, if you have ~/.profile, ~/.bash_profile or ~/.bash_login, add the commands there as well. If you have none of these, create a ~/.profile and add the commands there.

to add to ~/.profile:
echo 'export PYENV_ROOT="$HOME/.pyenv"' >> ~/.profile
echo '[[ -d $PYENV_ROOT/bin ]] && export PATH="$PYENV_ROOT/bin:$PATH"' >> ~/.profile
echo 'eval "$(pyenv init - bash)"' >> ~/.profile

to add to ~/.bash_profile:
echo 'export PYENV_ROOT="$HOME/.pyenv"' >> ~/.bash_profile
echo '[[ -d $PYENV_ROOT/bin ]] && export PATH="$PYENV_ROOT/bin:$PATH"' >> ~/.bash_profile
echo 'eval "$(pyenv init - bash)"' >> ~/.bash_profile

restart shell
exec "$SHELL"

安装python 3.11.11到~.pyenv/versions/3.11.11:
pyenv install 3.11.11

建立虚拟环境:
~/.pyenv/versions/3.11.11/bin/python -m venv ~/env/llm

4.安装open-webui
pip install open-webui
open-webui serve
可以用localhost:8080

4.1 本地部署(隐私优先)
使用你的 Ollama 实例作为 OpenAI 兼容端点:

import openai

# Connect to your local Ollama instance
client = openai.Client(
base_url="http://localhost:11434/v1",
api_key="ollama" # Authentication-free private access
)

response = client.chat.completions.create(
model="deepseek-r1:XXb", # change the "XX" by the distilled model you choose
messages=[{"role": "user", "content": "Explain blockchain security"}],
temperature=0.7 # Controls creativity vs precision
)
4.2 使用官方 DeepSeek-R1 云 API
对于可扩展的应用程序,请使用 DeepSeek 的官方 API,你可以在此处通过创建帐户并生成一个密钥来获取 DeepSeek API 密钥:

import openai
from dotenv import load_dotenv
import os

load_dotenv()
client = openai.OpenAI(
base_url="https://api.deepseek.com/v1",
api_key=os.getenv("DEEPSEEK_API_KEY")
)

response = client.chat.completions.create(
model="deepseek-reasoner",
messages=[{"role": "user", "content": "Write web scraping code with error handling"}],
max_tokens=1000 # Limit costs for long responses
)

posted on   medsci  阅读(222)  评论(0编辑  收藏  举报
相关博文:
阅读排行:
· 【.NET】调用本地 Deepseek 模型
· CSnakes vs Python.NET:高效嵌入与灵活互通的跨语言方案对比
· DeepSeek “源神”启动!「GitHub 热点速览」
· 我与微信审核的“相爱相杀”看个人小程序副业
· Plotly.NET 一个为 .NET 打造的强大开源交互式图表库
< 2025年2月 >
26 27 28 29 30 31 1
2 3 4 5 6 7 8
9 10 11 12 13 14 15
16 17 18 19 20 21 22
23 24 25 26 27 28 1
2 3 4 5 6 7 8

点击右上角即可分享
微信分享提示