How to Deploy Deepseek on Bivocom TG465

How to Deploy Deepseek on Bivocom TG465

DeepSeek is a specialized deep learning model developed by Meta AI. It is designed for information retrieval and data analysis tasks, particularly useful in handling large datasets commonly found in IoT applications.

DeepSeek Capabilities:

Information Retrieval: DeepSeek can efficiently search and extract specific information from vast datasets, enabling quick access to relevant data.
Data Analysis and Summarization: It can analyze complex datasets to identify patterns, trends, and anomalies. Additionally, DeepSeek is adept at summarizing extensive information into concise and meaningful summaries.
Content Generation: The model is capable of generating new content based on the patterns and structures it has learned from the data it has been trained on.

Relationship between Ollama and DeepSeek:

Ollama: Ollama and DeepSeek are both variants of the same underlying language model (like Chatgpt), developed by Meta AI for different purposes.
DeepSeek: Specializes in data analysis and information retrieval within IoT contexts.

This document will tell you how to run DeepSeek and test on TG465. In order to deploy DeepSeek R1 locally, we need to install Ollama first.

1. Environment and Related Tools Deployment

  • Install Ollama

curl -L https://ollama.com/download/ollama-linux-arm64.tgz -o ollama-linux-arm64.tgz
sudo tar -C /usr -xzf ollama-linux-arm64.tgz

Install Ollama1
  •  Create a New User and Group for Ollama

sudo useradd -r -s /bin/false -U -m -d /usr/share/ollama ollama
sudo usermod -a -G ollama $(whoami)

  • Add the service configuration to enable Ollama to be managed by Systemd, /etc/systemd/system/ollama.service:

(1) Enter path /etc/systemd/system/, create ollama.service file.
cd /etc/systemd/system/
touch ollama.service

(2) vi ollama.service
[Unit]
Description=Ollama Service
After=network-online.target

[Service]
ExecStart=/usr/bin/ollama serve
User=ollama
Group=ollama
Restart=always
RestartSec=3
Environment=”PATH=$PATH”

[Install]
WantedBy=default.target

  •  Start Ollama Service

sudo systemctl daemon-reload
sudo systemctl enable ollama

Start Ollama Service4

2. Model Download and Launch

  • Start Ollama and Check Running Status

sudo systemctl start ollama
sudo systemctl status ollama

Start Ollama and Check Running Status5
  •  Download and Run deepseek-r1-1.5b Model

ollama run deepseek-r1:1.5b

ollama run deepseek-r1 1.5b6
  •  Test

test7
Print Email

Comment

Be the first one who leave the comment.

Leave a Reply

We use cookies to deliver you the best experience. By browsing our website you agree to our use of cookies.