Ollama
This document was translated from Chinese by AI and has not yet been reviewed.
Ollama
Ollama is an excellent open-source tool that allows you to easily run and manage various Large Language Models (LLMs) locally. Cherry Studio now supports Ollama integration, enabling you to interact directly with locally deployed LLMs in a familiar interface, without relying on cloud services!
What is Ollama?
Ollama is a tool that simplifies the deployment and use of Large Language Models (LLMs). It has the following features:
Local Execution: Models run entirely on your local computer, without needing an internet connection, protecting your privacy and data security.
Easy to Use: Download, run, and manage various LLMs with simple command-line instructions.
Rich Model Library: Supports many popular open-source models like Llama 2, Deepseek, Mistral, and Gemma.
Cross-Platform: Supports macOS, Windows, and Linux systems.
Open API: Supports an OpenAI-compatible interface, allowing integration with other tools.
Why Use Ollama in Cherry Studio?
No Cloud Services Needed: No longer limited by cloud API quotas and fees. Enjoy the full power of local LLMs.
Data Privacy: All your conversation data remains on your local machine, eliminating concerns about privacy leaks.
Offline Availability: Continue interacting with LLMs even without an internet connection.
Customization: Choose and configure the LLMs that best suit your needs.
Configuring Ollama in Cherry Studio
1. Install and Run Ollama
First, you need to install and run Ollama on your computer. Follow these steps:
Download Ollama: Visit the official Ollama website (https://ollama.com/) and download the appropriate installer for your operating system. On Linux, you can install Ollama directly with the following command:
Install Ollama: Follow the installer's instructions to complete the installation.
Download a Model: Open your terminal (or command prompt) and use the
ollama run
command to download the model you want to use. For example, to download the Llama 3.2 model, run:Ollama will automatically download and run the model.
Keep Ollama Running: Ensure that Ollama remains running while you are interacting with Ollama models through Cherry Studio.
2. Add the Ollama Provider in Cherry Studio
Next, add Ollama as a custom AI provider in Cherry Studio:
Open Settings: In the left navigation bar of the Cherry Studio interface, click on "Settings" (the gear icon).
Go to Model Services: On the settings page, select the "Model Services" tab.
Add Provider: Click on Ollama in the list.

3. Configure the Ollama Provider
Find the newly added Ollama in the provider list and configure its details:
Enable Status:
Ensure the switch on the far right of the Ollama provider is turned on, indicating it is enabled.
API Key:
Ollama does not require an API key by default. You can leave this field blank or fill it with any content.
API Endpoint:
Enter the local API address provided by Ollama. Typically, the address is:
If you have changed the port, please modify it accordingly.
Keep-Alive Time: This option sets the session keep-alive duration in minutes. If there are no new conversations within the set time, Cherry Studio will automatically disconnect from Ollama to release resources.
Model Management:
Click the "+ Add" button to manually add the names of the models you have already downloaded in Ollama.
For example, if you have already downloaded the
llama3.2
model usingollama run llama3.2
, you can enterllama3.2
here.Click the "Manage" button to edit or delete the added models.
Getting Started
Once the configuration is complete, you can select the Ollama provider and your downloaded model in the Cherry Studio chat interface to start conversing with your local LLM!
Tips and Tricks
First Model Run: When running a model for the first time, Ollama needs to download the model file, which may take some time. Please be patient.
View Available Models: Run the
ollama list
command in the terminal to see a list of the Ollama models you have downloaded.Hardware Requirements: Running large language models requires certain computing resources (CPU, memory, GPU). Please ensure your computer's configuration meets the model's requirements.
Ollama Documentation: You can click the
View Ollama documentation and models
link on the configuration page to quickly navigate to the official Ollama documentation.
最后更新于
这有帮助吗?