Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
This document was translated from Chinese by AI and has not yet been reviewed.
When the assistant does not have a default assistant model configured, the model selected by default for new conversations is the one set here. The model used for prompt optimization and word-selection assistant is also configured in this section.
After each conversation, a model is called to generate a topic name for the dialog. The model set here is the one used for naming.
The translation feature in input boxes for conversations, painting, etc., and the translation model in the translation interface all use the model set here.
The model used for the quick assistant feature. For details, see Quick Assistant.
This document was translated from Chinese by AI and has not yet been reviewed.
The drawing feature currently supports painting models from DMXAPI, TokenFlux, AiHubMix, and SiliconFlow. You can register an account at SiliconFlow and add it as a provider to use this feature.
For questions about parameters, hover your mouse over the ?
icon in corresponding areas to view descriptions.
This document was translated from Chinese by AI and has not yet been reviewed.
Quick Assistant is a convenient tool provided by Cherry Studio that allows you to quickly access AI functions in any application, enabling instant operations like asking questions, translation, summarization, and explanations.
Open Settings: Navigate to Settings
-> Shortcuts
-> Quick Assistant
.
Enable Switch: Find and toggle on the Quick Assistant
button.
Set Shortcut Key (Optional):
Default shortcut for Windows: Ctrl + E
Default shortcut for macOS: ⌘ + E
Customize your shortcut here to avoid conflicts or match your usage habits.
Activate: Press your configured shortcut key (or default shortcut) in any application to open Quick Assistant.
Interact: Within the Quick Assistant window, you can directly perform:
Quick Questions: Ask any question to the AI.
Text Translation: Input text to be translated.
Content Summarization: Input long text for summarization.
Explanation: Input concepts or terms requiring explanation.
Close: Press ESC or click anywhere outside the Quick Assistant window.
Shortcut Conflicts: Modify shortcuts if defaults conflict with other applications.
Explore More Functions: Beyond documented features, Quick Assistant may support operations like code generation and style transfer. Continuously explore during usage.
Feedback & Improvements: Report issues or suggestions to the Cherry Studio team via .
macOS 版本安装教程
This document was translated from Chinese by AI and has not yet been reviewed.
First, visit the official download page to download the Mac version, or click the direct link below
Please download the chip-specific version matching your Mac
After downloading, click here
Drag the icon to install
Find the Cherry Studio icon in Launchpad and click it. If the Cherry Studio main interface opens, the installation is successful.
This document was translated from Chinese by AI and has not yet been reviewed.
This interface allows for local and cloud data backup and recovery, local data directory inquiry and cache clearing, export settings, and third-party connections.
Currently, data backup supports three methods: local backup, WebDAV backup, and S3-compatible storage (object storage) backup. For specific introductions and tutorials, please refer to the following documents:
Export settings allow you to configure the export options displayed in the export menu, as well as set the default path for Markdown exports, display styles, and more.
Third-party connections allow you to configure Cherry Studio's connection with third-party applications for quickly exporting conversation content to your familiar knowledge management applications. Currently supported applications include: Notion, Obsidian, SiYuan Note, Yuque, Joplin. For specific configuration tutorials, please refer to the following documents:
This document was translated from Chinese by AI and has not yet been reviewed.
The Agents page is an assistant marketplace where you can select or search for desired model presets. Click on a card to add the assistant to your conversation page's assistant list.
You can also edit and create your own assistants on this page.
Click My
, then Create Agent
to start building your own assistant.
Windows 版本安装教程
This document was translated from Chinese by AI and has not yet been reviewed.
Note: Cherry Studio does not support installation on Windows 7.
Click to download and select the appropriate version
If the browser warns that the file is not trusted, simply choose to keep it
Keep
→Trust Cherry-Studio
This document was translated from Chinese by AI and has not yet been reviewed.
Note: Windows 7 is not supported for installing Cherry Studio.
This document was translated from Chinese by AI and has not yet been reviewed.
Key translation notes:
Preserved all Markdown formatting (headings, checkboxes)
Technical terms remain unchanged: "JavaScript", "SSO", "iOS", "Android"
Action descriptions standardized: "Quick Pop-up" for 快捷弹窗, "multi-model" for 多模型
Functional translations: "划词翻译" → "text selection translation"
Feature localization: "AI 通话" → "AI calls"
Maintained present tense for consistency
Preserved special characters and list formatting
Translated bracket content while keeping technical references (JavaScript)
Proper noun capitalization: "AI Notes"
This document was translated from Chinese by AI and has not yet been reviewed.
This page only introduces the interface features. For configuration tutorials, please refer to the Provider Configuration tutorial in the Basic Tutorials.
In Cherry Studio, a single provider supports multiple keys for round-robin usage, with the polling method being sequential from front to back.
Add multiple keys separated by English commas. For example:
sk-xxxx1,sk-xxxx2,sk-xxxx3,sk-xxxx4
Must use English commas.
When using built-in providers, it's generally not necessary to fill in the API address. If modification is needed, strictly follow the address provided in the official documentation.
If the provider gives an address in the format https://xxx.xxx.com/v1/chat/completions, only fill in the root address part (https://xxx.xxx.com).
Cherry Studio will automatically append the remaining path (/v1/chat/completions). Failure to comply may result in failure to function properly.
Generally, clicking the Manage
button at the bottom left of the provider configuration page will automatically fetch all supported models. Click +
in the fetched list to add models to the model list.
Click the check button next to the API Key input box to test whether the configuration is successful.
After successful configuration, be sure to turn on the switch in the upper right corner. Otherwise the provider will remain disabled and you won't find corresponding models in the model list.
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio's translation feature provides you with fast and accurate text translation services, supporting mutual translation between multiple languages.
The translation interface mainly consists of the following components:
Source Language Selection Area:
Any Language: Cherry Studio will automatically identify the source language and perform translation.
Target Language Selection Area:
Dropdown Menu: Select the language you wish to translate the text into.
Settings Button:
Clicking will jump to .
Scroll Synchronization:
Toggle to enable scroll sync (scrolling in either side will synchronize the other).
Text Input Box (Left):
Input or paste the text you need to translate.
Translation Result Box (Right):
Displays the translated text.
Copy Button: Click to copy the translation result to clipboard.
Translate Button:
Click this button to start translation.
Translation History (Top Left):
Click to view translation history records.
Select Target Language:
Choose your desired translation language in the Target Language Selection Area.
Input or Paste Text:
Enter or paste the text to be translated in the left text input box.
Start Translation:
Click the Translate
button.
View and Copy Results:
Translation results will appear in the right result box.
Click the copy button to save the result to clipboard.
Q: What to do about inaccurate translations?
A: While AI translation is powerful, it's not perfect. For professional fields or complex contexts, manual proofreading is recommended. You may also try switching different models.
Q: Which languages are supported?
A: Cherry Studio translation supports multiple major languages. Refer to Cherry Studio's official website or in-app instructions for the specific supported languages list.
Q: Can entire files be translated?
A: The current interface primarily handles text translation. For document translation, please use Cherry Studio's conversation page to add files for translation.
Q: How to handle slow translation speeds?
A: Translation speed may be affected by network connection, text length, or server load. Ensure stable network connectivity and be patient.
This document was translated from Chinese by AI and has not yet been reviewed.
Email [email protected] to obtain editing privileges
Subject: Request for Cherry Studio Docs Editing Privileges
Body: State your reason for applying
This document was translated from Chinese by AI and has not yet been reviewed.
Note: Gemini image generation must be used in the chat interface because Gemini performs multi-modal interactive image generation and does not support parameter adjustment.
This document was translated from Chinese by AI and has not yet been reviewed.
Follow our social accounts: Twitter (X), Xiaohongshu, Weibo, Bilibili, Douyin
Join our community: QQ Group (575014769), Telegram, Discord, WeChat Group (Click to view)
Cherry Studio is an all-in-one AI assistant platform integrating multi-model conversations, knowledge base management, AI painting, translation, and more. Cherry Studio's highly customizable design, powerful extensibility, and user-friendly experience make it an ideal choice for both professional users and AI enthusiasts. Whether you are a beginner or a developer, you can find suitable AI features in Cherry Studio to improve work efficiency and creativity.
Multi-model Responses: Supports generating responses to the same question simultaneously from multiple models, allowing users to compare the performance of different models. For details, see Chat Interface.
Automatic Grouping: Conversation records for each assistant are automatically grouped for easy access to historical conversations.
Conversation Export: Supports exporting full or partial conversations to various formats (e.g., Markdown, Word), convenient for storage and sharing.
Highly Customizable Parameters: In addition to basic parameter adjustments, it also supports users filling in custom parameters to meet personalized needs.
Assistant Marketplace: Built-in over a thousand industry-specific assistants covering translation, programming, writing, and more, also supporting user-defined assistants.
Multi-format Rendering: Supports Markdown rendering, formula rendering, real-time HTML preview, and other features to enhance content display.
AI Painting: Provides a dedicated painting panel, allowing users to generate high-quality images through natural language descriptions.
AI Mini-Apps: Integrates various free web-based AI tools, allowing direct use without switching browsers.
Translation Function: Supports various translation scenarios, including dedicated translation panels, conversation translation, and prompt translation.
File Management: Files in conversations, paintings, and knowledge bases are uniformly categorized and managed to avoid tedious searching.
Global Search: Supports quickly locating historical records and knowledge base content, improving work efficiency.
Provider Model Aggregation: Supports unified invocation of models from mainstream providers such as OpenAI, Gemini, Anthropic, Azure.
Automatic Model Retrieval: One-click retrieval of the complete model list, no manual configuration required.
Multi-key Rotation: Supports rotating multiple API keys to avoid rate limit issues.
Accurate Avatar Matching: Automatically matches exclusive avatars for each model to improve recognition.
Custom Providers: Supports the integration of third-party providers that comply with OpenAI, Gemini, Anthropic, and other specifications, offering strong compatibility.
Custom CSS: Supports global style customization to create a unique interface style.
Custom Conversation Layout: Supports list or bubble style layouts, and allows custom message styles (e.g., code snippet styles).
Custom Avatars: Supports setting personalized avatars for software and assistants.
Custom Sidebar Menu: Users can hide or sort sidebar features according to their needs to optimize the user experience.
Multi-format Support: Supports importing various file formats such as PDF, DOCX, PPTX, XLSX, TXT, MD.
Multiple Data Source Support: Supports local files, URLs, sitemaps, and even manually entered content as knowledge base sources.
Knowledge Base Export: Supports exporting processed knowledge bases for sharing with others.
Search Inspection Support: After importing the knowledge base, users can perform real-time retrieval tests to check processing results and segmentation effects.
Quick Q&A: Invoke a quick assistant in any scenario (e.g., WeChat, browser) to quickly get answers.
Quick Translate: Supports quick translation of words or text in other scenarios.
Content Summarization: Quickly summarizes long text content to improve information extraction efficiency.
Explanation: Explains unclear questions with one click, no complex prompts needed.
Multiple Backup Solutions: Supports local backup, WebDAV backup, and timed backup to ensure data security.
Data Security: Supports full local usage scenarios, combined with local large models, to avoid data leakage risks.
Beginner-Friendly: Cherry Studio is committed to lowering technical barriers, allowing beginners to quickly get started and focus on work, study, or creation.
Comprehensive Documentation: Provides detailed user documentation and a common issues handbook to help users quickly resolve problems.
Continuous Iteration: The project team actively responds to user feedback and continuously optimizes features to ensure healthy project development.
Open Source and Extensibility: Supports users in customizing and extending through open-source code to meet personalized needs.
Knowledge Management and Query: Quickly build and query exclusive knowledge bases through the local knowledge base feature, suitable for research, education, and other fields.
Multi-model Conversation and Creation: Supports simultaneous multi-model conversations, helping users quickly obtain information or generate content.
Translation and Office Automation: Built-in translation assistant and file processing features, suitable for users needing cross-language communication or document processing.
AI Painting and Design: Generates images through natural language descriptions, meeting creative design needs.
This document was translated from Chinese by AI and has not yet been reviewed.
Log in to Alibaba Cloud Bailian. If you don't have an Alibaba Cloud account, you'll need to register.
Click the Create My API-KEY
button in the upper-right corner.
In the popup window, select the default workspace (or customize it if desired). You can optionally add a description.
Click the Confirm
button in the lower-right corner.
You should now see a new entry in the list. Click the View
button on the right.
Click the Copy
button.
Go to Cherry Studio, navigate to Settings
→ Model Services
→ Alibaba Cloud Bailian
, and paste the copied API key into the API Key
field.
You can adjust related settings as described in Model Services, then start using the service.
This document was translated from Chinese by AI and has not yet been reviewed.
On the official API Key page, click + Create new secret key
Copy the generated key, then open CherryStudio's Vendor Settings
Find the OpenAI vendor and enter the key you just obtained
Click Manage or Add at the bottom, add supported models, and toggle the vendor switch at the top right to start using.
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio data backup supports backup via S3 compatible storage (object storage). Common S3 compatible storage services include: AWS S3, Cloudflare R2, Alibaba Cloud OSS, Tencent Cloud COS, and MinIO.
Based on S3 compatible storage, multi-terminal data synchronization can be achieved by the method of Computer A
S3 Storage
Computer B
.
Create an object storage bucket, and record the bucket name. It is highly recommended to set the bucket to private read/write to prevent backup data leakage!!
Refer to the documentation, go to the cloud service console to obtain information such as Access Key ID
, Secret Access Key
, Endpoint
, Bucket
, Region
for S3 compatible storage.
Endpoint: The access address for S3 compatible storage, usually in the form of https://<bucket-name>.<region>.amazonaws.com
or https://<ACCOUNT_ID>.r2.cloudflarestorage.com
.
Region: The region where the bucket is located, such as us-west-1
, ap-southeast-1
, etc. For Cloudflare R2, please fill in auto
.
Bucket: The bucket name.
Access Key ID and Secret Access Key: Credentials used for authentication.
Root Path: Optional, specifies the root path when backing up to the bucket, default is empty.
Related Documentation
Cloudflare R2: Obtain Access Key ID and Secret Access Key
Alibaba Cloud OSS: Obtain Access Key ID and Access Key Secret
Tencent Cloud COS: Obtain SecretId and SecretKey
Fill in the above information in the S3 backup settings, click the backup button to perform backup, and click the manage button to view and manage the list of backup files.
This document was translated from Chinese by AI and has not yet been reviewed.
Data added to the Cherry Studio knowledge base is entirely stored locally. During the addition process, a copy of the document will be placed in the Cherry Studio data storage directory.
Vector Database: https://turso.tech/libsql
After documents are added to the Cherry Studio knowledge base, they are segmented into multiple fragments. These fragments are then processed by the embedding model.
When using large models for Q&A, relevant text fragments matching the query will be retrieved and processed together by the large language model.
If you have data privacy requirements, it is recommended to use local embedding databases and local large language models.
This document was translated from Chinese by AI and has not yet been reviewed.
Before obtaining the Gemini API key, you need a Google Cloud project (skip this step if you already have one).
Go to to create a project, fill in the project name, and click "Create Project".
On the official , click Create API Key
.
Copy the generated key, then open CherryStudio's .
Find the Gemini service provider and enter the key you just obtained.
Click "Manage" or "Add" at the bottom, add the supported models, then toggle the service provider switch at the top right to start using.
暂时不支持Claude模型
This document was translated from Chinese by AI and has not yet been reviewed.
Before obtaining the Gemini API Key, you need to have a Google Cloud project (if you already have one, this process can be skipped)
Go to to create a project, fill in the project name and click Create Project
Access the
Enable in the created project
Open the permissions page and create a service account
On the service account management page, find the service account you just created, click Keys
and create a new JSON key
After successful creation, the key file will be automatically saved to your computer in JSON format. Please store it securely
Select Vertex AI as the service provider
Fill in the corresponding fields from the JSON file
Click , and you can start using it!
This document was translated from Chinese by AI and has not yet been reviewed.
Use this method to clear CSS settings when incorrect CSS has been applied, or when you cannot access the settings interface after applying CSS.
Open the console by clicking on the CherryStudio window and pressing Ctrl+Shift+I (MacOS: command+option+I).
In the opened console window, click Console
Then manually enter document.getElementById('user-defined-custom-css').remove()
. Copying and pasting will most likely not work.
After entering, press Enter to confirm and clear the CSS settings. Then, go back to CherryStudio's display settings and remove the problematic CSS code.
This document was translated from Chinese by AI and has not yet been reviewed.
To use GitHub Copilot, you must first have a GitHub account and subscribe to the GitHub Copilot service. The free subscription tier is acceptable, but note that it does not support the latest Claude 3.7 model. For details, refer to the .
Click "Log in with GitHub" to generate and copy your Device Code.
After obtaining your Device Code, click the link to open your browser. Log in to your GitHub account, enter the Device Code, and grant authorization.
After successful authorization, return to Cherry Studio and click "Connect GitHub". Your GitHub username and avatar will appear upon successful connection.
Click the "Manage" button below, which will automatically fetch the currently supported models list online.
The current implementation uses Axios for network requests. Note that Axios does not support SOCKS proxies. Please use a system proxy or HTTP proxy, or avoid setting proxies within CherryStudio and use a global proxy instead. First, ensure your network connection is stable to prevent Device Code retrieval failures.
This document was translated from Chinese by AI and has not yet been reviewed.
Knowledge base document preprocessing requires upgrading Cherry Studio to v1.4.8 or higher.
After clicking 'Get API KEY', the application address will open in your browser. Click 'Apply Now' to fill out the form and obtain the API KEY, then fill it into the API KEY field.
Configure as shown above in the created knowledge base to complete the knowledge base document preprocessing configuration.
You can check knowledge base results by searching in the top right corner.
Knowledge Base Usage Tips: When using more capable models, you can change the knowledge base search mode to intent recognition. Intent recognition can describe your questions more accurately and broadly.
This document was translated from Chinese by AI and has not yet been reviewed.
1.2 Click the settings at the bottom left, and select 【SiliconFlow】 in the model service
1.2 Click the link to get SiliconCloud API Key
Log in to (if not registered, the first login will automatically register an account)
Visit to create a new key or copy an existing one
1.3 Click Manage to add models
Click the "Chat" button in the left menu bar
Enter text in the input box to start chatting
You can switch models by selecting the model name in the top menu
This document was translated from Chinese by AI and has not yet been reviewed.
MCP (Model Context Protocol) is an open-source protocol designed to provide context information to large language models (LLMs) in a standardized manner. For more details about MCP, see .
The following demonstrates how to use MCP in Cherry Studio using the fetch
feature as an example. Detailed information can be found in the .
Cherry Studio currently only uses built-in and , and will not reuse uv or bun already installed on your system.
In Settings > MCP Server
, click the Install
button to automatically download and install them. Downloads are sourced directly from GitHub and may be slow with a high probability of failure. Installation success should be verified by checking for files in the directories mentioned below.
Executable Installation Directories:
Windows: C:\Users\username\.cherrystudio\bin
macOS/Linux: ~/.cherrystudio/bin
If unable to install normally:
You can create symlinks from the system's corresponding commands to these locations. If the directory doesn't exist, create it manually. Alternatively, manually download executables and place them in this directory:
Bun: UV:
This document was translated from Chinese by AI and has not yet been reviewed.
Open Cherry Studio settings.
Find the MCP Server
option.
Click Add Server
.
Fill in the parameters for the MCP Server (). This may include:
Name: Custom name, e.g., fetch-server
Type: Select STDIO
Command: Enter uvx
Arguments: Enter mcp-server-fetch
(Additional parameters may be required depending on the specific Server)
Click Save
.
After completing the above configuration, Cherry Studio will automatically download the required MCP Server - fetch server
. Once downloaded, we can start using it! Note: If mcp-server-fetch configuration fails, try restarting your computer.
After successfully adding the MCP server in MCP Server
settings
As shown above, after integrating MCP's fetch
functionality, Cherry Studio can better understand user query intentions. It retrieves relevant information from the web to provide more accurate and comprehensive responses.
This document was translated from Chinese by AI and has not yet been reviewed.
Contact Person: Mr. Wang 📮: [email protected] 📱: 18954281942 (Not a customer service hotline)
For usage inquiries: • Join our user community at the bottom of the official website homepage • Email [email protected] • Or submit issues:
For additional guidance, join our
Commercial license details:
mcp-server-time
--local-timezone
<Your standard timezone, e.g., Asia/Shanghai>
This document was translated from Chinese by AI and has not yet been reviewed.
Automatically install MCP service (Beta)
Persistence memory base implementation based on local knowledge graph. This enables models to remember user-related information across different conversations.
MEMORY_FILE_PATH=/path/to/your/file.json
An MCP server implementation providing tools for dynamic and reflective problem-solving through structured thought processes.
An MCP server implementation integrated with Brave Search API, offering dual functionality for both web and local searches.
BRAVE_API_KEY=YOUR_API_KEY
MCP server for retrieving web content from URLs.
Node.js server implementing the Model Context Protocol (MCP) for file system operations.
This document was translated from Chinese by AI and has not yet been reviewed.
Join the Telegram discussion group for assistance: https://t.me/CherryStudioAI
GitHub Issues: https://github.com/CherryHQ/cherry-studio/issues/new/choose
Email the developers: [email protected]
This document was translated from Chinese by AI and has not yet been reviewed.
We welcome contributions to Cherry Studio! You can contribute in the following ways:
Contribute code: Develop new features or optimize existing code.
Fix bugs: Submit bug fixes you discover.
Maintain issues: Help manage GitHub issues.
Product design: Participate in design discussions.
Write documentation: Improve user manuals and guides.
Community engagement: Join discussions and assist users.
Promote usage: Spread the word about Cherry Studio.
Email [email protected]
Email subject: Application to Become a Developer Email content: Reason for Application
This document was translated from Chinese by AI and has not yet been reviewed.
360gpt-pro
8k
-
Not Supported
Dialogue
360AI_360gpt
360's flagship trillion-parameter large model, widely applicable to complex tasks across various domains.
glm-4v-flash
2k
1k
Not Supported
Dialogue, Visual Understanding
Zhipu_glm
Free model with robust image comprehension capabilities.
This document was translated from Chinese by AI and has not yet been reviewed.
Welcome to Cherry Studio (hereinafter referred to as "this software" or "we"). We highly value your privacy protection, and this Privacy Policy explains how we handle and protect your personal information and data. Please read and understand this agreement carefully before using this software:
To optimize user experience and improve software quality, we may only collect the following anonymous non-personal information: • Software version information; • Function activity and usage frequency; • Anonymous crash and error log information;
The above information is completely anonymous, does not involve any personally identifiable data, and cannot be associated with your personal information.
To maximize the protection of your privacy, we explicitly commit: • Will not collect, store, transmit, or process the model service API Key information you input into this software; • Will not collect, store, transmit, or process any conversation data generated during your use of this software, including but not limited to chat content, instruction information, knowledge base information, vector data, and other custom content; • Will not collect, store, transmit, or process any personally identifiable sensitive information.
This software uses the API Key of third-party model service providers that you apply for and configure yourself to complete model invocation and conversation functions. The model services you use (such as large language models, API interfaces, etc.) are provided and fully managed by the third-party provider you choose, and Cherry Studio only serves as a local tool providing interface invocation functionality with third-party model services.
Therefore: • All conversation data generated between you and the model service is unrelated to Cherry Studio. We neither participate in data storage nor perform any form of data transmission or relay; • You need to independently review and accept the privacy policies and relevant regulations of the corresponding third-party model service providers. These services' privacy policies can be viewed on each provider's official website.
You shall bear any privacy risks that may arise from using third-party model service providers. For specific privacy policies, data security measures, and related responsibilities, please refer to the relevant content on the official websites of your chosen model service providers. We assume no responsibility for these matters.
This agreement may be appropriately adjusted with software version updates. Please check it periodically. When substantive changes occur to the agreement, we will notify you through appropriate means.
If you have any questions regarding this agreement or Cherry Studio's privacy protection measures, please feel free to contact us at any time.
Thank you for choosing and trusting Cherry Studio. We will continue to provide you with a secure and reliable product experience.
This document was translated from Chinese by AI and has not yet been reviewed.
Have you ever experienced: saving 26 insightful articles in WeChat that you never open again, storing over 10 scattered files in a "Study Materials" folder on your computer, or trying to recall a theory you read half a year ago only to remember fragmented keywords? And when daily information exceeds your brain's processing limit, 90% of valuable knowledge gets forgotten within 72 hours. Now, by leveraging the Infini-AI large model service platform API + Cherry Studio to build a personal knowledge base, you can transform dust-collecting WeChat articles and fragmented course content into structured knowledge for precise retrieval.
1. Infini-AI API Service: The Stable "Thinking Hub" of Knowledge Bases
Serving as the "thinking hub" of knowledge bases, the Infini-AI large model service platform offers robust API services including full-capacity DeepSeek R1 and other model versions. Currently free to use without barriers after registration. Supports mainstream embedding models (bge, jina) for knowledge base construction. The platform continuously updates with the latest, most powerful open-source models, covering multiple modalities like images, videos, and audio.
2. Cherry Studio: Zero-Code Knowledge Base Setup
Compared to RAG knowledge base development requiring 1-2 months deployment time, Cherry Studio offers a significant advantage: zero-code operation. Instantly import multiple formats like Markdown/PDF/web pages – parsing 40MB files in 1 minute. Easily add local folders, saved WeChat articles, and course notes.
Step 1: Basic Preparation
Download the suitable version from Cherry Studio official site (https://cherry-ai.com/)
Register account: Log in to Infini-AI platform (https://cloud.infini-ai.com/genstudio/model?cherrystudio)
Get API key: Select "deepseek-r1" in Model Square, create and copy the API key + model name
Step 2: CherryStudio Settings
Go to model services → Select Infini-AI → Enter API key → Activate Infini-AI service
After setup, select the large model during interaction to use Infini-AI API in CherryStudio. Optional: Set as default model for convenience
Step 3: Add Knowledge Base
Choose any version of bge-series or jina-series embedding models from Infini-AI platform
After importing study materials, query: "Outline core formula derivations in Chapter 3 of 'Machine Learning'"
Result Demonstration
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio supports importing topics into Notion databases.
Visit Notion Integrations to create an application
Create an application
Name: Cherry Studio Type: Select the first option Icon: You can save this image
Copy the secret key and enter it in Cherry Studio settings
Open Notion website and create a new page. Select database type below, name it Cherry Studio, and connect as shown
If your Notion database URL looks like this:
https://www.notion.so/<long_hash_1>?v=<long_hash_2>
Then the Notion database ID is the part <long_hash_1>
Fill in Page Title Field Name
:
If your interface is in English, enter Name
If your interface is in Chinese, enter 名称
Congratulations! Notion configuration is complete ✅ You can now export Cherry Studio content to your Notion database
如何注册tavily?
This document was translated from Chinese by AI and has not yet been reviewed.
Visit the official website above, or go to Cherry Studio > Settings > Web Search > click "Get API Key" to directly access Tavily's login/registration page.
First-time users must register (Sign up) before logging in (Log in). The default page is the login interface.
Click "Sign up" and enter your email (or use Google/GitHub account) followed by your password.
🚨🚨🚨[Critical Step] After registration, a dynamic verification code is required. Scan the QR code to generate a one-time code.
Two solutions:
Download Microsoft Authenticator app (slightly complex)
Use WeChat Mini Program: Tencent Authenticator (recommended, as simple as it gets).
Search "Tencent Authenticator" in WeChat Mini Programs:
After completing these steps, you'll see the dashboard. Copy the API key to Cherry Studio to start using Tavily!
This document was translated from Chinese by AI and has not yet been reviewed.
Log in to Volcano Engine
Or click here to go directly
Click API Key Management in the sidebar
Create an API Key
After creation, click the eye icon next to the created API Key to reveal and copy it
Paste the copied API Key into Cherry Studio and then toggle the provider switch to ON
Enable the models you need at the bottom of the sidebar in the Ark console under Enablement Management. You can enable the Doubao series, DeepSeek, and other models as required
In the Model List Documentation, find the model ID corresponding to the desired model
Open Cherry Studio's Model Services settings and locate Volcano Engine
Click Add, then paste the previously obtained model ID into the model ID text field
Follow this process to add models one by one
There are two ways to write the API address:
First, the default in the client: https://ark.cn-beijing.volces.com/api/v3/
Second: https://ark.cn-beijing.volces.com/api/v3/chat/completions#
This document was translated from Chinese by AI and has not yet been reviewed.
Supports exporting topics and messages to SiYuan Note.
Open SiYuan Note and create a notebook
Open notebook settings and copy the Notebook ID
Paste the copied notebook ID into Cherry Studio settings
Fill in the SiYuan Note address
Local
Typically http://127.0.0.1:6806
Self-hosted
Use your domain http://note.domain.com
Copy the SiYuan Note API Token
Paste it into Cherry Studio settings and check
Congratulations, the configuration for SiYuan Note is complete ✅ Now you can export content from Cherry Studio to your SiYuan Note
This document was translated from Chinese by AI and has not yet been reviewed.
Create an account and log in at Huawei Cloud
Click this link to enter the MaaS control panel
Authorization
Click Authentication Management in the sidebar, create an API Key (secret key) and copy it
Then create a new service provider in CherryStudio
After creation, enter the secret key
Click Model Deployment in the sidebar, claim all offerings
Click Call
Copy the address from ①, paste it into CherryStudio's service provider address field and add a "#" at the end and add a "#" at the end and add a "#" at the end and add a "#" at the end and add a "#" at the end Why add "#"? [see here](https://docs.cherry-ai.com/cherrystudio/preview/settings/providers#api-di-zhi) > You can choose not to read that and just follow this tutorial; > Alternatively, you can fill it by removing v1/chat/completions - feel free to use your own method if you know how, but if not, strictly follow this tutorial.
Then copy the model name from ②, and click the "+ Add" button in CherryStudio to create a new model
Enter the model name exactly as shown - do not add or remove anything, and don't include quotes. Copy exactly as in the example.
Click the Add Model button to complete.
This document was translated from Chinese by AI and has not yet been reviewed.
Dify Knowledge Base MCP requires upgrading Cherry Studio to v1.2.9 or higher.
Open Search MCP
.
Add the dify-knowledge
server.
Requires configuring parameters and environment variables
Dify Knowledge Base key can be obtained in the following way:
This document was translated from Chinese by AI and has not yet been reviewed.
Log in and navigate to the tokens page
Create a new token (you can also directly use the default token ↑)
Copy the token
Open CherryStudio's service provider settings and click Add
at the bottom of the provider list
Enter a note name, select OpenAI as the provider, and click OK
Paste the key you just copied
Return to the API Key page, copy the root address from the browser's address bar, for example:
Add models (click Manage to auto-fetch or enter manually) and toggle the switch in the top right corner to start using.
The interface may differ in other OneAPI themes, but the addition method follows the same workflow as above.
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio data backup supports WebDAV for backup. You can choose a suitable WebDAV service for cloud backup.
Based on WebDAV, you can achieve multi-device data synchronization through Computer A
WebDAV
Computer B
.
Log in to Jianguoyun, click your username in the top right corner, and select "Account Info":
Select "Security Options" and click "Add Application":
Enter the application name and generate a random password;
Copy and record the password;
Get the server address, account, and password;
In Cherry Studio Settings -> Data Settings, fill in the WebDAV information;
Choose to back up or restore data, and you can set the automatic backup time period.
WebDAV services with a lower barrier to entry are generally cloud drives:
(requires membership)
(requires purchase)
(Free storage capacity is 10GB, single file size limit is 250MB.)
(Dropbox offers 2GB free, can expand by 16GB by inviting friends.)
(Free space is 10GB, an additional 5GB can be obtained through invitation.)
(Free users get 10GB capacity.)
Secondly, there are some services that require self-deployment:
This document was translated from Chinese by AI and has not yet been reviewed.
In version 0.9.1, Cherry Studio introduces the long-awaited knowledge base feature. Below is the step-by-step guide to using Cherry Studio's knowledge base.
Find models in the Model Management Service. You can quickly filter by clicking "Embedding Models".
Locate the required model and add it to My Models.
Access: Click the Knowledge Base icon in Cherry Studio's left toolbar to enter the management page.
Add Knowledge Base: Click "Add" to start creating.
Naming: Enter a name for the knowledge base and add an embedding model (e.g., bge-m3) to complete creation.
Add Files: Click the "Add Files" button to open file selection.
Select Files: Choose supported file formats (pdf, docx, pptx, xlsx, txt, md, mdx, etc.) and open.
Vectorization: The system automatically vectorizes files. A green checkmark (✓) indicates completion.
Cherry Studio supports adding data through:
Folder Directories: Entire directories where supported format files are auto-vectorized.
URL Links: e.g., https://docs.siliconflow.cn/introduction
Sitemaps: XML format sitemaps, e.g., https://docs.siliconflow.cn/sitemap.xml
Plain Text Notes: Custom content input.
After vectorization:
Click "Search Knowledge Base" at bottom of page.
Enter query content.
View search results.
Matching scores are displayed per result.
Create new topic > Click "Knowledge Base" in toolbar > Select target knowledge base.
Enter question > Model generates answer using retrieved knowledge.
Data sources appear below answer for quick reference.
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio is a free and open-source project. As the project grows, the workload of the team increases significantly. To reduce communication costs and efficiently resolve your issues, we encourage everyone to follow the steps below when reporting problems. This will allow our team to dedicate more time to project maintenance and development. Thank you for your cooperation!
Most basic problems can be resolved by carefully reviewing the documentation:
Functionality and usage questions can be answered in the documentation;
High-frequency issues are compiled on the page—check there first for solutions;
Complex questions can often be resolved through search or using the search bar;
Carefully read all hint boxes in documents to avoid many common issues;
Check existing solutions in the GitHub .
For model-related issues unrelated to client functionality (e.g., model errors, unexpected responses, parameter settings):
Search online for solutions first;
Provide error messages to AI assistants for resolution suggestions.
If steps 1-2 don't solve your problem:
Seek help in our official , , or () When reporting:
For model errors:
Provide full screenshots with error messages visible
Include console errors ()
Sensitive information can be redacted, but keep model names, parameters, and error details
For software bugs:
Give detailed error descriptions
Provide precise reproduction
Include OS (Windows/Mac/Linux) and software version number
For intermittent issues, describe scenarios and configurations comprehensively
Request Documentation or Suggest Improvements
Contact via Telegram @Wangmouuu
, QQ (1355873789
), or email [email protected]
.
This document was translated from Chinese by AI and has not yet been reviewed.
Requires upgrading Cherry Studio to v1.2.9 or higher.
In v1.2.9, Cherry Studio has officially partnered with ModelScope to simplify the process of adding MCP servers, avoiding configuration errors while enabling access to a vast number of MCP servers in the ModelScope community. Follow these steps to synchronize ModelScope's MCP servers in Cherry Studio.
Go to MCP Server Settings in Settings and select Sync Servers
Select ModelScope and browse available MCP services
Register/login to ModelScope and view MCP service details
In the MCP service details, choose "Connect Service"
Click "Get API" token in Cherry Studio to visit ModelScope's official website, copy the API token, then paste it back in Cherry Studio.
ModelScope-connected MCP services will appear in Cherry Studio's MCP server list and become available for conversations.
For future MCP servers connected via ModelScope website, simply click Sync Servers
to add them incrementally.
With these steps, you've mastered how to efficiently synchronize ModelScope's MCP servers in Cherry Studio. The simplified configuration process eliminates manual setup complexities while unlocking access to ModelScope's extensive MCP server resources.
Start exploring these powerful MCP services to enhance your Cherry Studio experience!
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio not only integrates mainstream AI model services but also empowers you with powerful customization capabilities. Through the Custom AI Providers feature, you can easily integrate any AI model you require.
Flexibility: Break free from predefined provider lists and freely choose the AI models that best suit your needs.
Diversity: Experiment with various AI models from different platforms to discover their unique advantages.
Controllability: Directly manage your API keys and access addresses to ensure security and privacy.
Customization: Integrate privately deployed models to meet the demands of specific business scenarios.
Add your custom AI provider to Cherry Studio in just a few simple steps:
Open Settings: Click the "Settings" (gear icon) in the left navigation bar of the Cherry Studio interface.
Enter Model Services: Select the "Model Services" tab in the settings page.
Add Provider: On the "Model Services" page, you'll see existing providers. Click the "+ Add" button below the list to open the "Add Provider" pop-up.
Fill in Information: In the pop-up, provide the following details:
Provider Name: Give your custom provider a recognizable name (e.g., MyCustomOpenAI).
Provider Type: Select your provider type from the dropdown menu. Currently supports:
OpenAI
Gemini
Anthropic
Azure OpenAI
Save Configuration: After filling in the details, click the "Add" button to save your configuration.
After adding, locate your newly added provider in the list and configure it:
Activation Status: Toggle the activation switch on the far right of the list to enable this custom service.
API Key:
Enter the API key provided by your AI provider.
Click the "Test" button to verify the key's validity.
API Address:
Enter the base URL to access the AI service.
Always refer to your AI provider's official documentation for the correct API address.
Model Management:
Click the "+ Add" button to manually add model IDs you want to use under this provider (e.g., gpt-3.5-turbo
, gemini-pro
).
If unsure about specific model names, consult your AI provider's official documentation.
Click the "Manage" button to edit or delete added models.
After completing the above configurations, you can select your custom AI provider and model in Cherry Studio's chat interface and start conversing with AI!
vLLM is a fast and easy-to-use LLM inference library similar to Ollama. Here's how to integrate vLLM into Cherry Studio:
Install vLLM: Follow vLLM's official documentation (https://docs.vllm.ai/en/latest/getting_started/quickstart.html) to install vLLM.
pip install vllm # if using pip
uv pip install vllm # if using uv
Launch vLLM Service: Start the service using vLLM's OpenAI-compatible interface via two main methods:
Using vllm.entrypoints.openai.api_server
python -m vllm.entrypoints.openai.api_server --model gpt2
Using uvicorn
vllm --model gpt2 --served-model-name gpt2
Ensure the service launches successfully, listening on the default port 8000
. You can also specify a different port using the --port
parameter.
Add vLLM Provider in Cherry Studio:
Follow the steps above to add a new custom AI provider.
Provider Name: vLLM
Provider Type: Select OpenAI
.
Configure vLLM Provider:
API Key: Leave this field blank or enter any value since vLLM doesn't require an API key.
API Address: Enter vLLM's API address (default: http://localhost:8000/
, adjust if using a different port).
Model Management: Add the model name loaded in vLLM (e.g., gpt2
for the command python -m vllm.entrypoints.openai.api_server --model gpt2
).
Start Chatting: Now select the vLLM provider and the gpt2
model in Cherry Studio to chat with the vLLM-powered LLM!
Read Documentation Carefully: Before adding custom providers, thoroughly review your AI provider's official documentation for API keys, addresses, model names, etc.
Test API Keys: Use the "Test" button to quickly verify API key validity.
Verify API Addresses: Different providers and models may have varying API addresses—ensure correctness.
Add Models Judiciously: Only add models you'll actually use to avoid cluttering.
This document was translated from Chinese by AI and has not yet been reviewed.
Ollama is an excellent open-source tool that allows you to easily run and manage various large language models (LLMs) locally. Cherry Studio now supports Ollama integration, enabling you to interact directly with locally deployed LLMs through the familiar interface without relying on cloud services!
Ollama is a tool that simplifies the deployment and use of large language models (LLMs). It has the following features:
Local Operation: Models run entirely on your local computer without requiring internet connectivity, protecting your privacy and data security.
Simple and User-Friendly: Download, run, and manage various LLMs through simple command-line instructions.
Rich Model Selection: Supports popular open-source models like Llama 2, Deepseek, Mistral, Gemma, and more.
Cross-Platform: Compatible with macOS, Windows, and Linux systems.
OpenAPI: Features OpenAI-compatible interfaces for integration with other tools.
No Cloud Service Needed: Break free from cloud API quotas and fees, and fully leverage the power of local LLMs.
Data Privacy: All your conversation data remains locally stored with no privacy concerns.
Offline Availability: Continue interacting with LLMs even without an internet connection.
Customization: Choose and configure the most suitable LLMs according to your needs.
First, you need to install and run Ollama on your computer. Follow these steps:
Download Ollama: Visit the Ollama official website and download the installation package for your operating system. For Linux systems, you can directly install Ollama by running:
curl -fsSL https://ollama.com/install.sh | sh
Install Ollama: Complete the installation by following the installer instructions.
Download Models: Open a terminal (or command prompt) and use the ollama run
command to download your desired model. For example, to download the Llama 2 model, run:
ollama run llama3.2
Ollama will automatically download and run the model.
Keep Ollama Running: Ensure Ollama remains running while using Cherry Studio to interact with its models.
Next, add Ollama as a custom AI provider in Cherry Studio:
Open Settings: Click on "Settings" (gear icon) in the left navigation bar of Cherry Studio.
Access Model Services: Select the "Model Services" tab in the settings page.
Add Provider: Click "Ollama" in the provider list.
Locate the newly added Ollama provider in the list and configure it in detail:
Enable Status:
Ensure the switch on the far right of the Ollama provider is turned on (indicating enabled status).
API Key:
Ollama typically requires no API key. Leave this field blank or enter any content.
API Address:
Enter Ollama's local API address. Normally, this is:
http://localhost:11434/
Adjust if you've modified the default port.
Keep-Alive Time: Sets the session retention time in minutes. Cherry Studio will automatically disconnect from Ollama if no new interactions occur within this period to free up resources.
Model Management:
Click "+ Add" to manually add the names of models you've downloaded in Ollama.
For example, if you downloaded llama3.2
via ollama run llama3.2
, enter llama3.2
here.
Click "Manage" to edit or delete added models.
After completing the above configurations, select Ollama as the provider and choose your downloaded model in Cherry Studio's chat interface to start conversations with local LLMs!
First-Time Model Execution: Ollama needs to download model files during initial runs, which may take considerable time. Please be patient.
View Available Models: Run ollama list
in the terminal to see your downloaded Ollama models.
Hardware Requirements: Running large language models requires substantial computing resources (CPU, RAM, GPU). Ensure your computer meets the model's requirements.
Ollama Documentation: Click the View Ollama Documentation and Models
link in the configuration page to quickly access Ollama's official documentation.
如何在 Cherry Studio 使用联网模式
This document was translated from Chinese by AI and has not yet been reviewed.
In the Cherry Studio question window, click the 【Globe】 icon to enable internet access.
Mode 1: Models with built-in internet function from providers
When using such models, enabling internet access requires no extra steps - it's straightforward.
Quickly identify internet-enabled models by checking for a small globe icon next to the model name above the chat interface.
This method also helps quickly distinguish internet-enabled models in the Model Management page.
Cherry Studio currently supports internet-enabled models from
Google Gemini
OpenRouter (all models support internet)
Tencent Hunyuan
Zhipu AI
Alibaba Bailian, etc.
Important note:
Special cases exist where models may access internet without the globe icon, as explained in the tutorial below.
Mode 2: Models without internet function use Tavily service
When using models without built-in internet (no globe icon), use Tavily search service to process real-time information.
First-time Tavily setup triggers a setup prompt - simply follow the instructions!
After clicking, you'll be redirected to Tavily's website to register/login. Create and copy your API key back to Cherry Studio.
Registration guide available in Tavily tutorial within this documentation directory.
Tavily registration reference:
The following interface confirms successful registration.
Test again for results: shows normal internet search with default result count (5).
Note: Tavily has monthly free tier limits - exceeding requires payment~~
PS: Please report any issues you encounter.
This document was translated from Chinese by AI and has not yet been reviewed.
On this page, you can set the software's color theme, page layout, or use Custom CSS for personalized configurations.
Here you can set the default interface color mode (Light mode, Dark mode, or Follow System).
These settings apply to the layout of the conversation interface.
Conversation Panel Position
Auto-Switch to Conversation
When enabled, clicking an assistant name will automatically switch to the corresponding conversation.
Show Conversation Time
When enabled, displays the conversation creation time below the conversation.
This setting allows flexible customization of the interface. Refer to the advanced tutorial on Custom CSS for specific methods.
This document was translated from Chinese by AI and has not yet been reviewed.
Automatic installation of MCP requires upgrading Cherry Studio to v1.1.18 or higher.
In addition to manual installation, Cherry Studio has a built-in @mcpmarket/mcp-auto-install
tool that provides a more convenient way to install MCP servers. You simply need to enter specific commands in conversations with large models that support MCP services.
Beta Phase Reminder:
@mcpmarket/mcp-auto-install
is currently in beta phase
Effectiveness depends on the "intelligence" of the large model - some configurations are automatically added, while others still require manual parameter adjustments in MCP settings
Current search sources are from @modelcontextprotocol, but this can be customized (explained below)
For example, you can input:
Help me install a filesystem mcp server
The system will automatically recognize your requirements and complete the installation via @mcpmarket/mcp-auto-install
. This tool supports various types of MCP servers, including but not limited to:
filesystem (file system)
fetch (network requests)
sqlite (database)
etc.
The MCP_PACKAGE_SCOPES variable allows customization of MCP service search sources. Default value:
@modelcontextprotocol
.
@mcpmarket/mcp-auto-install
Library数据设置→Obsidian配置
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio supports integration with Obsidian, allowing you to export complete conversations or individual messages to your Obsidian vault.
This process does not require installing additional Obsidian plugins. However, since the mechanism used by Cherry Studio to import to Obsidian is similar to the Obsidian Web Clipper, it is recommended to upgrade Obsidian to the latest version (the current version should be at least greater than 1.7.2) to prevent .
Open Cherry Studio's Settings → Data Settings → Obsidian Configuration menu. The drop-down box will automatically display the names of Obsidian vaults that have been opened on this machine. Select your target Obsidian vault:
Exporting a Complete Conversation
Return to Cherry Studio's conversation interface, right-click on the conversation, select Export, and click Export to Obsidian:
A window will pop up allowing you to configure the Properties of the note exported to Obsidian, its folder location, and the processing method:
Vault: Click the drop-down menu to select another Obsidian vault
Path: Click the drop-down menu to select the folder for storing the exported conversation note
As Obsidian note properties:
Tags
Created time
Source
Three available processing methods:
New (overwrite if exists): Create a new conversation note in the folder
specified at Path. If a note with the same name exists, it will overwrite the old note
Prepend: When a note with the same name already exists, export the selected conversation content and add it to the beginning of that note
Append: When a note with the same name already exists, export the selected conversation content and add it to the end of that note
After selecting all options, click OK to export the complete conversation to the specified folder in the corresponding Obsidian vault.
Exporting a Single Message
To export a single message, click the three-dash menu below the message, select Export, and click Export to Obsidian:
The same window as when exporting a complete conversation will appear, requiring you to configure the note properties and processing method. Complete the configuration following the same steps as in .
🎉 Congratulations! You've completed all configurations for Cherry Studio's integration with Obsidian and finished the entire export process. Enjoy yourselves!
Open your Obsidian vault and create a folder
to store exported conversations (shown as "Cherry Studio" folder in the example):
Note the text highlighted in the bottom-left corner - this is your vault
name.
In Cherry Studio's Settings → Data Settings → Obsidian Configuration menu, enter the vault
name and folder
name obtained in :
Global Tags
is optional and can be used to set tags for all exported conversations in Obsidian. Fill in as needed.
Exporting a Complete Conversation
Return to Cherry Studio's conversation interface, right-click on the conversation, select Export, and click Export to Obsidian:
A window will pop up allowing you to adjust the Properties of the note exported to Obsidian and select a processing method. Three processing methods are available:
New (overwrite if exists): Create a new conversation note in the folder
specified in . If a note with the same name exists, it will overwrite the old note
Prepend: When a note with the same name already exists, export the selected conversation content and add it to the beginning of that note
Append: When a note with the same name already exists, export the selected conversation content and add it to the end of that note
Exporting a Single Message
To export a single message, click the three-dash menu below the message, select Export, and click Export to Obsidian:
The same window as when exporting a complete conversation will appear. Complete the configuration by following the same steps in .
🎉 Congratulations! You've completed all configurations for Cherry Studio's integration with Obsidian and finished the entire export process. Enjoy yourselves!
This document was translated from Chinese by AI and has not yet been reviewed.
By customizing CSS, you can modify the software's appearance to better suit your preferences, like this:
For more theme variables, refer to the source code:
Cherry Studio Theme Library:
Share some Chinese-style Cherry Studio theme skins:
This document was translated from Chinese by AI and has not yet been reviewed.
Tokens are the basic units of text processed by AI models, understood as the smallest units of model "thinking". They are not entirely equivalent to the characters or words as we understand them, but rather a special way the model segments text itself.
1. Chinese Tokenization
A Chinese character is typically encoded as 1-2 tokens
For example: "你好"
≈ 2-4 tokens
2. English Tokenization
Common words are typically 1 token
Longer or uncommon words are decomposed into multiple tokens
For example:
"hello"
= 1 token
"indescribable"
= 4 tokens
3. Special Characters
Spaces, punctuation, etc. also consume tokens
Line breaks are typically 1 token
A Tokenizer is the tool that converts text into tokens for AI models. It determines how to split input text into the smallest units that models can understand.
1. Different Training Data
Different corpora lead to different optimization directions
Variations in multilingual support levels
Specialized optimization for specific domains (medical, legal, etc.)
2. Different Tokenization Algorithms
BPE (Byte Pair Encoding) - OpenAI GPT series
WordPiece - Google BERT
SentencePiece - Suitable for multilingual scenarios
3. Different Optimization Goals
Some focus on compression efficiency
Others on semantic preservation
Others on processing speed
The same text may have different token counts across models:
Basic Concept: An embedding model is a technique that converts high-dimensional discrete data (text, images, etc.) into low-dimensional continuous vectors. This transformation allows machines to better understand and process complex data. Imagine it as simplifying a complex puzzle into a simple coordinate point that still retains the puzzle's key features. In the large model ecosystem, it serves as a "translator," converting human-understandable information into AI-computable numerical forms.
Working Principle: Taking natural language processing as an example, an embedding model maps words to specific positions in vector space. In this space:
Vectors for "King" and "Queen" will be very close
Pet-related words like "cat" and "dog" will cluster together
Semantically unrelated words like "car" and "bread" will be distant
Main Application Scenarios:
Text analysis: Document classification, sentiment analysis
Recommendation systems: Personalized content recommendations
Image processing: Similar image retrieval
Search engines: Semantic search optimization
Core Advantages:
Dimensionality reduction: Simplifies complex data into manageable vector forms
Semantic preservation: Retains key semantic information from original data
Computational efficiency: Significantly improves training and inference efficiency
Technical Value: Embedding models are fundamental components of modern AI systems, providing high-quality data representations for machine learning tasks, and are key technologies driving advances in natural language processing, computer vision, and other fields.
Basic Workflow:
Knowledge Base Preprocessing Stage
Split documents into appropriately sized chunks
Use embedding models to convert each chunk into vectors
Store vectors and original text in a vector database
Query Processing Stage
Convert user questions into vectors
Retrieve similar content from the vector database
Provide retrieved context to the LLM
MCP is an open-source protocol designed to provide context information to large language models (LLMs) in a standardized way.
Analogy: Think of MCP as a "USB drive" for AI. Just as USB drives can store various files and be plugged into computers for immediate use, MCP servers can "plug in" various context-providing "plugins". LLMs can request these plugins from MCP servers as needed to obtain richer context information and enhance their capabilities.
Comparison with Function Tools: Traditional function tools provide external capabilities to LLMs, but MCP is a higher-level abstraction. Function tools focus on specific tasks, while MCP provides a more universal, modular context acquisition mechanism.
Standardization: Provides unified interfaces and data formats for seamless collaboration between LLMs and context providers.
Modularity: Allows developers to decompose context information into independent modules (plugins) for easy management and reuse.
Flexibility: Enables LLMs to dynamically select required context plugins for smarter, more personalized interactions.
Extensibility: Supports adding new types of context plugins in the future, providing unlimited possibilities for LLM capability expansion.
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio supports two methods for configuring blacklists: manual setup and adding subscription sources. Configuration rules reference
You can add rules to search results or click the toolbar icon to block specified websites. Rules can be specified using: (example: *://*.example.com/*
) or (example: /example\.(net|org)/
).
You can also subscribe to public rule sets. This website lists some subscriptions: https://iorate.github.io/ublacklist/subscriptions
Here are some recommended subscription source links:
Input: "Hello, world!"
GPT-3: 4 tokens
BERT: 3 tokens
Claude: 3 tokens
https://git.io/ublacklist
Chinese
https://raw.githubusercontent.com/laylavish/uBlockOrigin-HUGE-AI-Blocklist/main/list_uBlacklist.txt
AI Generated
This document was translated from Chinese by AI and has not yet been reviewed.
By using or distributing any portion or element of Cherry Studio materials, you are deemed to have acknowledged and accepted the terms of this Agreement, which shall take effect immediately.
This Cherry Studio License Agreement (hereinafter referred to as the "Agreement") shall mean the terms and conditions governing the use, reproduction, distribution, and modification of the Materials as defined herein.
"We" (or "Our") shall mean Shanghai WisdomAI Technology Co., Ltd.
"You" (or "Your") shall mean a natural person or legal entity exercising rights granted under this Agreement and/or using the Materials for any purpose and in any field of use.
"Third Party" shall mean an individual or legal entity not under common control with Us or You.
"Cherry Studio" shall mean this software suite, including but not limited to [e.g., core libraries, editor, plugins, sample projects], as well as source code, documentation, sample code, and other elements of the foregoing distributed by Us. (Please elaborate based on Cherry Studio’s actual composition.)
"Materials" shall collectively refer to Cherry Studio and documentation (and any portions thereof), proprietary to Shanghai WisdomAI Technology Co., Ltd., and provided under this Agreement.
"Source" Form shall mean the preferred form for making modifications, including but not limited to source code, documentation source files, and configuration files.
"Object" Form shall mean any form resulting from mechanical transformation or translation of a Source Form, including but not limited to compiled object code, generated documentation, and forms converted into other media types.
"Commercial Use" shall mean use for the purpose of direct or indirect commercial benefit or commercial advantage, including but not limited to sale, licensing, subscription, advertising, marketing, training, consulting services, etc.
"Modification" shall mean any change, adaptation, derivation, or secondary development to the Source Form of the Materials, including but not limited to modifying application names, logos, code, features, interfaces, etc.
Free Commercial Use (Limited to Unmodified Code): Subject to the terms and conditions of this Agreement, We hereby grant You a non-exclusive, worldwide, non-transferable, royalty-free license to use, reproduce, distribute, copy, and disseminate unmodified Materials, including for Commercial Use, under the intellectual property or other rights We own or control embodied in the Materials.
Commercial Authorization (When Necessary): Under the conditions described in Section III ("Commercial Authorization"), You must obtain explicit written commercial authorization from Us to exercise rights under this Agreement.
In any of the following circumstances, You must contact Us and obtain explicit written commercial authorization before continuing to use Cherry Studio Materials:
Modification and Derivation: You modify Cherry Studio Materials or develop derivatives based on them (including but not limited to modifying application names, logos, code, features, interfaces, etc.).
Enterprise Services: You provide services based on Cherry Studio within Your enterprise or for enterprise clients, and such services support 10 or more cumulative users.
Hardware Bundling: You pre-install or integrate Cherry Studio into hardware devices or products for bundled sales.
Government or Educational Institutional Procurement: Your use case involves large-scale procurement projects by government or educational institutions, especially those involving sensitive requirements such as security or data privacy.
Public Cloud Services: Providing public-facing cloud services based on Cherry Studio.
You may distribute copies of unmodified Materials or make them available as part of a product or service containing unmodified Materials, distributed in Source or Object Form, provided You satisfy the following conditions:
You must provide a copy of this Agreement to any other recipient of the Materials;
You must retain the following attribution notice in all copies of the Materials You distribute, placing it within a "NOTICE" or similar text file distributed as part of such copies:
"Cherry Studio is licensed under the Cherry Studio LICENSE AGREEMENT, Copyright (c) Shanghai WisdomAI Technology Co., Ltd. All Rights Reserved."
Materials may be subject to export controls or restrictions. You shall comply with applicable laws and regulations when using the Materials.
If You use Materials or any output thereof to create, train, fine-tune, or improve software or models that will be distributed or provided, We encourage You to prominently mark "Built with Cherry Studio" or "Powered by Cherry Studio" in relevant product documentation.
We retain ownership of all intellectual property rights to the Materials and derivative works made by or for Us. Subject to compliance with the terms and conditions of this Agreement, intellectual property rights in modifications and derivative works of the Materials created by You shall be governed by the specific commercial authorization agreement. Without obtaining commercial authorization, You do not own such modifications or derivatives, and all intellectual property rights therein remain vested in Us.
No trademark license is granted for the use of Our trade names, trademarks, service marks, or product names except as necessary to fulfill notice obligations under this Agreement or for reasonable and customary use in describing and redistributing the Materials.
If You initiate litigation or other legal proceedings (including counterclaims in litigation) against Us or any entity alleging that the Materials, any output thereof, or any portion thereof infringes any intellectual property or other rights owned or licensable by You, all licenses granted to You under this Agreement shall terminate as of the commencement or filing date of such litigation or proceedings.
We have no obligation to support, update, provide training for, or develop any further versions of Cherry Studio Materials, nor to grant any related licenses.
THE MATERIALS ARE PROVIDED "AS IS" WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING WARRANTIES OF MERCHANTABILITY, NON-INFRINGEMENT, OR FITNESS FOR A PARTICULAR PURPOSE. WE MAKE NO WARRANTIES AND ASSUME NO LIABILITY REGARDING THE SECURITY OR STABILITY OF THE MATERIALS OR ANY OUTPUT THEREOF.
IN NO EVENT SHALL WE BE LIABLE TO YOU FOR ANY DAMAGES ARISING FROM YOUR USE OR INABILITY TO USE THE MATERIALS OR ANY OUTPUT THEREOF, INCLUDING BUT NOT LIMITED TO ANY DIRECT, INDIRECT, SPECIAL, OR CONSEQUENTIAL DAMAGES, HOWEVER CAUSED.
You shall defend, indemnify, and hold Us harmless from any claims by third parties arising from or related to Your use or distribution of the Materials.
The term of this Agreement shall begin upon Your acceptance hereof or access to the Materials and shall continue until terminated in accordance with these terms and conditions.
We may terminate this Agreement if You breach any term or condition herein. Upon termination, You must cease all use of the Materials. Sections VII, IX, and "II. Contributor Agreement" shall survive termination.
This Agreement and any disputes arising out of or in connection herewith shall be governed by the laws of China.
The People's Court of Shanghai shall have exclusive jurisdiction over any disputes arising from this Agreement.
:root {
--color-background: #1a462788;
--color-background-soft: #1a4627aa;
--color-background-mute: #1a462766;
--navbar-background: #1a4627;
--chat-background: #1a4627;
--chat-background-user: #28b561;
--chat-background-assistant: #1a462722;
}
#content-container {
background-color: #2e5d3a !important;
}
:root {
font-family: "汉仪唐美人" !important; /* Font */
}
/* Color of expanded deep thinking text */
.ant-collapse-content-box .markdown {
color: red;
}
/* Theme variables */
:root {
--color-black-soft: #2a2b2a; /* Dark background color */
--color-white-soft: #f8f7f2; /* Light background color */
}
/* Dark theme */
body[theme-mode="dark"] {
/* Colors */
--color-background: #2b2b2b; /* Dark background color */
--color-background-soft: #303030; /* Light background color */
--color-background-mute: #282c34; /* Neutral background color */
--navbar-background: var(-–color-black-soft); /* Navigation bar background */
--chat-background: var(–-color-black-soft); /* Chat background */
--chat-background-user: #323332; /* User chat background */
--chat-background-assistant: #2d2e2d; /* Assistant chat background */
}
/* Dark theme specific styles */
body[theme-mode="dark"] {
#content-container {
background-color: var(-–chat-background-assistant) !important; /* Content container background */
}
#content-container #messages {
background-color: var(-–chat-background-assistant); /* Messages background */
}
.inputbar-container {
background-color: #3d3d3a; /* Input bar background */
border: 1px solid #5e5d5940; /* Input bar border color */
border-radius: 8px; /* Input bar border radius */
}
/* Code styles */
code {
background-color: #e5e5e20d; /* Code background */
color: #ea928a; /* Code text color */
}
pre code {
color: #abb2bf; /* Preformatted code text color */
}
}
/* Light theme */
body[theme-mode="light"] {
/* Colors */
--color-white: #ffffff; /* White */
--color-background: #ebe8e2; /* Light background */
--color-background-soft: #cbc7be; /* Light background */
--color-background-mute: #e4e1d7; /* Neutral background */
--navbar-background: var(-–color-white-soft); /* Navigation bar background */
--chat-background: var(-–color-white-soft); /* Chat background */
--chat-background-user: #f8f7f2; /* User chat background */
--chat-background-assistant: #f6f4ec; /* Assistant chat background */
}
/* Light theme specific styles */
body[theme-mode="light"] {
#content-container {
background-color: var(-–chat-background-assistant) !important; /* Content container background */
}
#content-container #messages {
background-color: var(-–chat-background-assistant); /* Messages background */
}
.inputbar-container {
background-color: #ffffff; /* Input bar background */
border: 1px solid #87867f40; /* Input bar border color */
border-radius: 8px; /* Input bar border radius */
}
/* Code styles */
code {
background-color: #3d39290d; /* Code background */
color: #7c1b13; /* Code text color */
}
pre code {
color: #000000; /* Preformatted code text color */
}
}
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio data storage follows system specifications. Data is automatically placed in the user directory at the following locations:
macOS: /Users/username/Library/Application Support/CherryStudioDev
Windows: C:\Users\username\AppData\Roaming\CherryStudio
Linux: /home/username/.config/CherryStudio
You can also view it at:
Method 1:
This can be achieved by creating a symbolic link. Exit the application, move the data to your desired location, then create a link at the original location pointing to the new path.
For detailed steps, refer to: https://github.com/CherryHQ/cherry-studio/issues/621#issuecomment-2588652880
Method 2: Based on Electron application characteristics, modify the storage location by configuring launch parameters.
--user-data-dir Example: Cherry-Studio-*-x64-portable.exe --user-data-dir="%user_data_dir%"
Example:
PS D:\CherryStudio> dir
目录: D:\CherryStudio
Mode LastWriteTime Length Name
---- ------------- ------ ----
d----- 2025/4/18 14:05 user-data-dir
-a---- 2025/4/14 23:05 94987175 Cherry-Studio-1.2.4-x64-portable.exe
-a---- 2025/4/18 14:05 701 init_cherry_studio.bat
init_cherry_studio.bat (encoding: ANSI)
@title CherryStudio Initialization
@echo off
set current_path_dir=%~dp0
@echo Current path: %current_path_dir%
set user_data_dir=%current_path_dir%user-data-dir
@echo CherryStudio data path: %user_data_dir%
@echo Searching for Cherry-Studio-*-portable.exe in current directory
setlocal enabledelayedexpansion
for /f "delims=" %%F in ('dir /b /a-d "Cherry-Studio-*-portable*.exe" 2^>nul') do ( # Compatible with GitHub and official releases. Modify for other versions
set "target_file=!cd!\%%F"
goto :break
)
:break
if defined target_file (
echo Found file: %target_file%
) else (
echo No matching files found. Exiting script
pause
exit
)
@echo Press any key to continue...
pause
@echo Launching CherryStudio
start %target_file% --user-data-dir="%user_data_dir%"
@echo Operation completed
@echo on
exit
Initial structure of user-data-dir:
PS D:\CherryStudio> dir .\user-data-dir\
目录: D:\CherryStudio\user-data-dir
Mode LastWriteTime Length Name
---- ------------- ------ ----
d----- 2025/4/18 14:29 blob_storage
d----- 2025/4/18 14:07 Cache
d----- 2025/4/18 14:07 Code Cache
d----- 2025/4/18 14:07 Data
d----- 2025/4/18 14:07 DawnGraphiteCache
d----- 2025/4/18 14:07 DawnWebGPUCache
d----- 2025/4/18 14:07 Dictionaries
d----- 2025/4/18 14:07 GPUCache
d----- 2025/4/18 14:07 IndexedDB
d----- 2025/4/18 14:07 Local Storage
d----- 2025/4/18 14:07 logs
d----- 2025/4/18 14:30 Network
d----- 2025/4/18 14:07 Partitions
d----- 2025/4/18 14:29 Session Storage
d----- 2025/4/18 14:07 Shared Dictionary
d----- 2025/4/18 14:07 WebStorage
-a---- 2025/4/18 14:07 36 .updaterId
-a---- 2025/4/18 14:29 20 config.json
-a---- 2025/4/18 14:07 434 Local State
-a---- 2025/4/18 14:29 57 Preferences
-a---- 2025/4/18 14:09 4096 SharedStorage
-a---- 2025/4/18 14:30 140 window-state.json
This document was translated from Chinese by AI and has not yet been reviewed.
4xx (Client Error Status Codes): Generally indicate requests that cannot be completed due to syntax errors, authentication/authorization failures, etc.
5xx (Server Error Status Codes): Generally indicate server-side errors like server downtime, request timeouts, etc.
400
Request body format error, etc.
Check error message in chat response or use the to view error details, then follow prompts.
【Common Case 1】 For Gemini models, card binding may be required; 【Common Case 2】 Data size exceeds limit, common with vision models where image size exceeds provider's request limit; 【Common Case 3】 Added unsupported parameters or incorrect parameter values. Try testing with a clean assistant; 【Common Case 4】 Context exceeds limit - clear context, start new conversation, or reduce context messages.
401
Authentication failure: Unsupported model or provider account suspension
Contact service provider or check account status
403
No permission for requested operation
Follow instructions based on error message in chat response or
404
Requested resource not found
Check request paths, etc.
422
Semantically invalid request despite correct format
Common with JSON semantic errors (e.g., null values; numbers/booleans passed where strings required)
429
Request rate limit reached
Request rate (TPM/RPM) capped, try again later
500
Internal server error
Contact service provider if persistent
501
Server doesn't support requested functionality
502
Bad Gateway: Invalid response from upstream server
503
Service Unavailable: Server overloaded/maintenance
504
Gateway Timeout: Proxy server didn't receive timely response from upstream
Press Ctrl + Shift + I when focused on Cherry Studio client window (Mac: Command + Option + I)
In console window, click Network
→ locate last entry marked with red ×
labeled completions
(for dialogue/translation/model check errors) or generations
(for image generation errors) → click Response
to view full error message (area ④ in image).
If unable to diagnose error, screenshot this interface and share in official community
This method works for dialogue, model testing, knowledge base creation, image generation, etc. Always open console before making requests.
If formula code displays instead of rendering, check delimiters:
Delimiter Usage Inline Formulas
Single dollar sign:
$formula$
Or
\(formula\)
Formula Blocks
Double dollar sign:
$$formula$$
Or
\[formula\]
Example:
$$\sum_{i=1}^n x_i$$
For rendering errors/garbled text (common with Chinese in formulas), try switching rendering engine to KateX.
Model unavailable
Verify provider supports the model and check service status.
Used non-embedding model
First confirm if the model supports image recognition. Hot models in Cherry Studio are categorized - models with an eye icon after the name support image recognition.
Image-capable models support file uploads. If model features aren't matched correctly:
Go to provider's model list
Click settings icon next to model name
Enable image option
Check model details on provider's page. Vision-incompatible models won't benefit from enabling image options.
This document was translated from Chinese by AI and has not yet been reviewed.
Log in and open the token page
Click "Add Token"
Enter a token name and click Submit (configure other settings if needed)
Open CherryStudio's provider settings and click Add
at the bottom of the provider list
Enter a remark name, select OpenAI as the provider, and click OK
Paste the key you just copied
Return to the API Key acquisition page and copy the root address from the browser's address bar, for example:
Add models (click Manage to automatically fetch or manually enter) and toggle the switch at the top right to use.
This document was translated from Chinese by AI and has not yet been reviewed.
An Assistant
is a personalized configuration for the selected model, such as preset prompts and parameters. These settings help the model better align with your expected workflow.
The System Default Assistant
comes with relatively universal parameters (without prompts). You can use it directly or find the presets you need on the Agents page.
The Assistant
is the parent set of Topics
. A single assistant can create multiple topics (i.e., conversations). All Topics
share the parameter settings and preset prompts (prompt) of the Assistant
.
New Topic
Creates a new topic under the current assistant.
Upload Image or Document
Image upload requires model support. Uploading documents will automatically parse text as context for the model.
Web Search
Requires configuration of web search information in settings. Search results are provided as context to the LLM. See Web Search Mode.
Knowledge Base
Enables the knowledge base. See Knowledge Base Tutorial.
MCP Server
Enables MCP server functionality. See MCP Usage Tutorial.
Generate Image
Hidden by default. For models that support image generation (e.g., Gemini), manually enable this button to generate images.
Select Model
Switches to the specified model for subsequent conversations while preserving context.
Quick Phrases
Requires predefined common phrases in settings. Invoke them here for direct input, supporting variables.
Clear Messages
Deletes all content in this topic.
Expand
Enlarges the dialog box for long text input.
Clear Context
Truncates the context available to the model without deleting content—the model "forgets" previous conversation content.
Estimated Token Count
Shows estimated token usage: Current Context
, Max Context
(∞ indicates unlimited context), Current Message Word Count
, and Estimated Tokens
.
Translate
Translates the current input box content into English.
Model settings synchronize with the Model Settings
parameters in assistant settings. See Edit Assistant.
Message Separator
:
Uses a divider to separate message content from the action bar.
Use Serif Font
:
Switches font style. You can also change fonts via Custom CSS.
Show Line Numbers for Code
:
Displays line numbers in code blocks generated by the model.
Collapsible Code Blocks
:
Automatically collapses code blocks when code snippets are long.
Wrap Lines in Code Blocks
:
Enables automatic line wrapping when single-line code exceeds window width.
Auto-Collapse Reasoning Content
:
Automatically collapses reasoning processes for models that support step-by-step thinking.
Message Style
:
Switches chat interface to bubble style or list style.
Code Style
:
Changes display style for code snippets.
Math Formula Engine
:
KaTeX: Faster rendering with performance optimization
MathJax: Slower rendering with comprehensive symbol and command support
Message Font Size
:
Adjusts font size in the chat interface.
Show Estimated Token Count
:
Displays estimated token consumption for input text in the input box (reference only, not actual context tokens).
Paste Long Text as File
:
Long text pasted into the input box automatically appears as files to reduce input interference.
Render Input Messages with Markdown
:
When disabled, only renders model responses, not sent messages.
Triple-Space Translation
:
Tap spacebar three times to translate input content to English after typing a message.
Note: This action overwrites the original text.
Target Language
:
Sets target language for both translation button and triple-space translation.
In the assistant interface, select the assistant name → choose corresponding settings in the right-click menu
Prompt Settings
Name
:
Customizable assistant name for easy identification.
Prompt
:
i.e., prompt. Edit content following prompt writing examples on the Agents page.
Model Settings
Default Model
:
Sets a fixed default model for the assistant. When adding from Agents page or copying assistant, initial model uses this setting. If unset, initial model = global default model (see Default Assistant Model).
Auto-Reset Model
:
When enabled: After switching models during conversation, creating a new topic resets to assistant's default model. When disabled: New topics inherit the previous topic's model.
Example: Assistant default model = gpt-3.5-turbo. Create Topic 1 → switch to gpt-4o during conversation.
Enabled Auto-Reset: Topic 2 uses gpt-3.5-turbo
Disabled Auto-Reset: Topic 2 uses gpt-4o
Temperature
:
Controls randomness/creativity of text generation (default=0.7):
Low (0-0.3): More deterministic output. Ideal for code generation, data analysis
Medium (0.4-0.7): Balanced creativity/coherence. Recommended for chatbots (~0.5)
High (0.8-1.0): High creativity/diversity. Ideal for creative writing, but reduces coherence
Top P (Nucleus Sampling)
:
Default=1. Lower values → more focused/comprehensible responses. Higher values → wider vocabulary diversity.
Sampling controls token probability thresholds:
Low (0.1-0.3): Conservative output. Ideal for code comments/tech docs
Medium (0.4-0.6): Balanced diversity/accuracy. General dialogue/writing
High (0.7-1.0): Diverse expression. Creative writing scenarios
Context Window
Number of messages to retain in context. Higher values → longer context → higher token usage:
5-10: Normal conversations
>10: Complex tasks requiring longer memory (e.g., multi-step content generation)
Note: More messages = higher token consumption
Enable Message Length Limit (MaxToken)
Sets maximum tokens per response. Directly impacts answer quality/length.
Example: When testing model connectivity, set MaxToken=1 to confirm response without specific content.
Most models support up to 32k tokens (some 64k+—check model documentation).
Suggestions:
Normal chat: 500-800
Short text gen: 800-2000
Code gen: 2000-3600
Long text gen: 4000+ (requires model support)
Responses are truncated at MaxToken limit. Incomplete expressions or truncation (e.g., long code) may occur—adjust as needed.
Stream Output
Enables continuous data stream processing instead of batch transmission. Provides real-time response generation (typing effect) in clients like CherryStudio.
Disabled: Full response delivered at once (like WeChat messages)
Enabled: Character-by-character output (generates → transmits each token immediately)
Custom Parameters
Adds extra request parameters to the body (e.g., presence_penalty
). Generally not needed for regular use.
Parameters like top-p, max_tokens, and stream belong to this category.
Format: Parameter name—Parameter type (text/number/etc.)—Value. See documentation: Click here
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio is a multi-model desktop client currently supporting installation packages for Windows, Linux, and macOS systems. It integrates mainstream LLM models to provide multi-scenario assistance. Users can enhance work efficiency through smart conversation management, open-source customization, and multi-theme interfaces.
Cherry Studio is now deeply integrated with PPIO's High-Performance API Channel – leveraging enterprise-grade computing power to ensure high-speed response for DeepSeek-R1/V3 and 99.9% service availability, delivering a fast and smooth experience.
The tutorial below provides a complete integration solution (including API key configuration), enabling the advanced mode of Cherry Studio Intelligent Scheduling + PPIO High-Performance API within 3 minutes.
First download Cherry Studio from the official website: (If inaccessible, download your required version from Quark Netdisk: )
(1) Click the settings icon in the bottom left corner, set the provider name to PPIO
, and click "OK"
(2) Visit , click 【User Avatar】→【API Key Management】 to enter console
Click 【+ Create】 to generate a new API key. Customize a key name. Generated keys are visible only at creation – immediately copy and save them to avoid affecting future usage
(3) In Cherry Studio settings, select 【PPIO Paiou Cloud】, enter the API key generated on the official website, then click 【Verify】
(4) Select model: using deepseek/deepseek-r1/community as example. Switch directly if needing other models.
DeepSeek R1 and V3 community versions are for trial use only. They are full-parameter models with identical stability and performance. For high-volume usage, top up and switch to non-community versions.
(1) After clicking 【Verify】 and seeing successful connection, it's ready for use
(2) Finally, click 【@】 and select the newly added DeepSeek R1 model under PPIO provider to start chatting~
【Partial material source: 】
For visual learners, we've prepared a Bilibili video tutorial. Follow step-by-step instructions to quickly master PPIO API + Cherry Studio configuration. Click the link to jump directly: →
【Video material source: sola】
This document was translated from Chinese by AI and has not yet been reviewed.
Cherry Studio supports web search via SearXNG, an open-source project deployable both locally and on servers. Its configuration differs slightly from other setups requiring API providers.
SearXNG Project Link: SearXNG
Open-source and free, no API required
Relatively high privacy
Highly customizable
SearXNG doesn't require complex environment configuration and can deploy using Docker with just a single available port.
1. Download, install, and configure docker
After installation, select an image storage path:
2. Search and pull the SearXNG image
Enter searxng in the search bar:
Pull the image:
3. Run the image
After successful pull, go to Images:
Select the pulled image and click run:
Open settings for configuration:
Take port 8085
as example:
After successful run, open SearXNG frontend via link:
This page indicates successful deployment:
Since Docker installation on Windows can be cumbersome, users can deploy SearXNG on servers for sharing. Unfortunately, SearXNG currently lacks built-in authentication, making deployed instances vulnerable to scanning and abuse.
To address this, Cherry Studio supports HTTP Basic Authentication (RFC7617). If exposing SearXNG publicly, you must configure HTTP Basic Authentication via reverse proxies like Nginx. This tutorial assumes basic Linux operations knowledge.
Similarly deploy via Docker. Assuming Docker CE is installed per official guide, run these commands on fresh Debian systems:
sudo apt update
sudo apt install git -y
# Pull official repo
cd /opt
git clone https://github.com/searxng/searxng-docker.git
cd /opt/searxng-docker
# Set to false for limited server bandwidth
export IMAGE_PROXY=true
# Modify config
cat <<EOF > /opt/searxng-docker/searxng/settings.yml
# see https://docs.searxng.org/admin/settings/settings.html#settings-use-default-settings
use_default_settings: true
server:
# base_url is defined in the SEARXNG_BASE_URL environment variable, see .env and docker-compose.yml
secret_key: $(openssl rand -hex 32)
limiter: false # can be disabled for a private instance
image_proxy: $IMAGE_PROXY
ui:
static_use_hash: true
redis:
url: redis://redis:6379/0
search:
formats:
- html
- json
EOF
Edit docker-compose.yaml
to change ports or reuse existing Nginx:
version: "3.7"
services:
# Remove below if reusing existing Nginx instead of Caddy
caddy:
container_name: caddy
image: docker.io/library/caddy:2-alpine
network_mode: host
restart: unless-stopped
volumes:
- ./Caddyfile:/etc/caddy/Caddyfile:ro
- caddy-data:/data:rw
- caddy-config:/config:rw
environment:
- SEARXNG_HOSTNAME=${SEARXNG_HOSTNAME:-http://localhost}
- SEARXNG_TLS=${LETSENCRYPT_EMAIL:-internal}
cap_drop:
- ALL
cap_add:
- NET_BIND_SERVICE
logging:
driver: "json-file"
options:
max-size: "1m"
max-file: "1"
# Remove above if reusing existing Nginx instead of Caddy
redis:
container_name: redis
image: docker.io/valkey/valkey:8-alpine
command: valkey-server --save 30 1 --loglevel warning
restart: unless-stopped
networks:
- searxng
volumes:
- valkey-data2:/data
cap_drop:
- ALL
cap_add:
- SETGID
- SETUID
- DAC_OVERRIDE
logging:
driver: "json-file"
options:
max-size: "1m"
max-file: "1"
searxng:
container_name: searxng
image: docker.io/searxng/searxng:latest
restart: unless-stopped
networks:
- searxng
# Default host port:8080, change to "127.0.0.1:8000:8080" for port 8000
ports:
- "127.0.0.1:8080:8080"
volumes:
- ./searxng:/etc/searxng:rw
environment:
- SEARXNG_BASE_URL=https://${SEARXNG_HOSTNAME:-localhost}/
- UWSGI_WORKERS=${SEARXNG_UWSGI_WORKERS:-4}
- UWSGI_THREADS=${SEARXNG_UWSGI_THREADS:-4}
cap_drop:
- ALL
cap_add:
- CHOWN
- SETGID
- SETUID
logging:
driver: "json-file"
options:
max-size: "1m"
max-file: "1"
networks:
searxng:
volumes:
# Remove below if reusing existing Nginx
caddy-data:
caddy-config:
# Remove above if reusing existing Nginx
valkey-data2:
Start with docker compose up -d
. View logs using docker compose logs -f searxng
.
For server panels like Baota or 1Panel:
Add site and configure Nginx reverse proxy per their docs
Locate Nginx config, modify per below example:
server
{
listen 443 ssl;
# Your hostname
server_name search.example.com;
# index index.html;
# root /data/www/default;
# SSL config
ssl_certificate /path/to/your/cert/fullchain.pem;
ssl_certificate_key /path/to/your/cert/privkey.pem;
# HSTS
# add_header Strict-Transport-Security "max-age=31536000; includeSubDomains; preload";
# Reverse proxy config
location / {
# Add these two lines in location block
auth_basic "Please enter your username and password";
auth_basic_user_file /etc/nginx/conf.d/search.htpasswd;
proxy_http_version 1.1;
proxy_set_header Connection "";
proxy_redirect off;
proxy_set_header Host $host;
proxy_set_header X-Forwarded-For $proxy_protocol_addr;
proxy_pass http://127.0.0.1:8000;
client_max_body_size 0;
}
# access_log ...;
# error_log ...;
}
Save password file in /etc/nginx/conf.d
:
echo "example_name:$(openssl passwd -5 'example_password')" > /etc/nginx/conf.d/search.htpasswd
Restart or reload Nginx. Access should prompt for credentials:
After successful SearXNG deployment:
Go to web search settings > Select Searxng
Initial verification may fail due to missing JSON format:
In Docker > Files tab, locate tagged folder > settings.yml
Edit file, add "json" to formats (line 78)
Rerun image
Successful verification in Cherry Studio
Use:
Local: http://localhost:port
Docker: http://host.docker.internal:port
For server deployments with HTTP Basic Authentication:
Initial verification returns 401
Configure credentials in client:
To customize search engines:
Default preferences don't affect model invocations
Configure model-invoked engines in settings.yml:
Syntax reference:
For lengthy edits, modify in local IDE then paste back.
Add "json" to return formats:
Cherry Studio defaults to engines with "web" and "general" categories (e.g., Google). In mainland China, force Baidu:
use_default_settings:
engines:
keep_only:
- baidu
engines:
- name: baidu
engine: baidu
categories:
- web
- general
disabled: false
Disable limiter in settings:
This document was translated from Chinese by AI and has not yet been reviewed.
Official Model Information Reference
Doubao-embedding
4095
Doubao-embedding-vision
8191
Doubao-embedding-large
4095
Official Model Information Reference
text-embedding-v3
8192
text-embedding-v2
2048
text-embedding-v1
2048
text-embedding-async-v2
2048
text-embedding-async-v1
2048
Official Model Information Reference
text-embedding-3-small
8191
text-embedding-3-large
8191
text-embedding-ada-002
8191
Official Model Information Reference
Embedding-V1
384
tao-8k
8192
Official Model Information Reference
embedding-2
1024
embedding-3
2048
Official Model Information Reference
hunyuan-embedding
1024
Official Model Information Reference
Baichuan-Text-Embedding
512
Official Model Information Reference
M2-BERT-80M-2K-Retrieval
2048
M2-BERT-80M-8K-Retrieval
8192
M2-BERT-80M-32K-Retrieval
32768
UAE-Large-v1
512
BGE-Large-EN-v1.5
512
BGE-Base-EN-v1.5
512
Official Model Information Reference
jina-embedding-b-en-v1
512
jina-embeddings-v2-base-en
8191
jina-embeddings-v2-base-zh
8191
jina-embeddings-v2-base-de
8191
jina-embeddings-v2-base-code
8191
jina-embeddings-v2-base-es
8191
jina-colbert-v1-en
8191
jina-reranker-v1-base-en
8191
jina-reranker-v1-turbo-en
8191
jina-reranker-v1-tiny-en
8191
jina-clip-v1
8191
jina-reranker-v2-base-multilingual
8191
reader-lm-1.5b
256000
reader-lm-0.5b
256000
jina-colbert-v2
8191
jina-embeddings-v3
8191
Official Model Information Reference
BAAI/bge-m3
8191
netease-youdao/bce-embedding-base_v1
512
BAAI/bge-large-zh-v1.5
512
BAAI/bge-large-en-v1.5
512
Pro/BAAI/bge-m3
8191
Official Model Information Reference
text-embedding-004
2048
Official Model Information Reference
nomic-embed-text-v1
8192
nomic-embed-text-v1.5
8192
gte-multilingual-base
8192
Official Model Information Reference
embedding-query
4000
embedding-passage
4000
Official Model Information Reference
embed-english-v3.0
512
embed-english-light-v3.0
512
embed-multilingual-v3.0
512
embed-multilingual-light-v3.0
512
embed-english-v2.0
512
embed-english-light-v2.0
512
embed-multilingual-v2.0
256
This document was translated from Chinese by AI and has not yet been reviewed.
This is a leaderboard based on Chatbot Arena (lmarena.ai) data, generated through an automated process.
Data Update Time: 2025-07-07 11:42:50 UTC / 2025-07-07 19:42:50 CST (Beijing Time)
1
1
1477
+5/-5
15,769
Proprietary
N/A
2
2
1446
+4/-5
13,997
Proprietary
N/A
3
3
1429
+4/-4
24,237
OpenAI
Proprietary
N/A
3
2
1427
+3/-4
21,965
OpenAI
Proprietary
N/A
3
6
1425
+4/-5
12,847
DeepSeek
MIT
N/A
3
7
1422
+3/-4
25,763
xAI
Proprietary
N/A
5
6
1418
+4/-4
21,209
Proprietary
N/A
6
4
1414
+5/-4
15,271
OpenAI
Proprietary
N/A
9
7
1398
+5/-5
17,002
Proprietary
N/A
9
11
1392
+5/-4
15,758
Alibaba
Apache 2.0
N/A
11
6
1384
+3/-4
18,275
OpenAI
Proprietary
N/A
11
12
1382
+3/-3
21,008
DeepSeek
MIT
N/A
11
17
1380
+6/-5
8,247
Tencent
Proprietary
N/A
11
11
1376
+6/-6
8,058
MiniMax
Apache 2.0
N/A
13
12
1374
+3/-5
19,430
DeepSeek
MIT
N/A
14
19
1370
+4/-4
19,980
Mistral
Proprietary
N/A
14
6
1370
+4/-4
20,056
Anthropic
Proprietary
N/A
15
23
1367
+4/-4
14,597
Alibaba
Apache 2.0
N/A
16
11
1366
+2/-3
29,038
OpenAI
Proprietary
N/A
16
11
1363
+4/-4
17,974
OpenAI
Proprietary
N/A
17
23
1363
+3/-3
32,074
Alibaba
Proprietary
N/A
18
25
1363
+3/-3
36,915
Proprietary
N/A
18
31
1359
+6/-5
10,561
xAI
Proprietary
N/A
19
25
1360
+3/-3
26,443
Gemma
N/A
24
31
1344
+12/-7
4,074
Alibaba
Apache 2.0
N/A
25
19
1351
+3/-4
33,177
OpenAI
Proprietary
2023/10
25
12
1343
+4/-5
16,050
Anthropic
Proprietary
N/A
26
25
1340
+4/-4
19,404
OpenAI
Proprietary
N/A
26
32
1337
+7/-8
3,976
Gemma
N/A
26
23
1337
+5/-4
17,292
OpenAI
Proprietary
N/A
26
31
1334
+4/-4
22,841
DeepSeek
DeepSeek
N/A
26
31
1332
+13/-13
2,061
Mistral
Apache 2.0
N/A
28
38
1333
+4/-5
18,386
Alibaba
Apache 2.0
N/A
29
36
1327
+8/-6
6,028
Zhipu
Proprietary
N/A
30
31
1329
+4/-4
26,104
Proprietary
N/A
30
56
1327
+5/-7
7,517
Amazon
Proprietary
N/A
30
32
1326
+7/-6
6,055
Alibaba
Proprietary
N/A
30
31
1321
+10/-11
2,656
Nvidia
Nvidia Open Model
N/A
32
32
1326
+3/-3
24,524
Cohere
CC-BY-NC-4.0
N/A
33
42
1323
+4/-4
14,229
Alibaba
Apache 2.0
N/A
33
38
1321
+7/-8
5,126
StepFun
Proprietary
N/A
33
31
1318
+8/-10
2,452
Tencent
Proprietary
N/A
34
39
1312
+11/-12
2,371
Nvidia
Nvidia
N/A
35
39
1320
+2/-2
54,951
OpenAI
Proprietary
2023/10
35
32
1319
+3/-3
36,971
OpenAI
Proprietary
N/A
38
32
1318
+2/-2
58,645
Proprietary
N/A
38
33
1312
+8/-10
2,510
Tencent
Proprietary
N/A
41
18
1313
+4/-4
25,955
Anthropic
Proprietary
N/A
43
58
1307
+7/-8
7,379
Gemma
N/A
44
21
1306
+4/-3
30,677
Anthropic
Proprietary
N/A
47
48
1304
+2/-2
67,084
xAI
Proprietary
2024/3
47
50
1303
+4/-3
28,968
01 AI
Proprietary
N/A
48
35
1301
+2/-2
117,747
OpenAI
Proprietary
2023/10
48
63
1298
+4/-6
10,715
Alibaba
Proprietary
N/A
50
25
1299
+2/-2
77,905
Anthropic
Proprietary
2024/4
50
55
1295
+6/-6
7,243
DeepSeek
DeepSeek
N/A
52
74
1292
+8/-9
4,321
Gemma
N/A
55
43
1292
+4/-4
18,010
Meta
Llama 4
N/A
55
64
1291
+3/-3
26,074
NexusFlow
NexusFlow
N/A
55
60
1290
+3/-3
27,788
Zhipu AI
Proprietary
N/A
55
49
1288
+8/-7
3,856
Tencent
Proprietary
N/A
55
56
1287
+6/-8
6,302
OpenAI
Proprietary
N/A
56
70
1287
+3/-3
37,021
Proprietary
N/A
56
79
1284
+5/-7
7,577
Nvidia
Llama 3.1
2023/12
57
61
1288
+2/-2
72,473
OpenAI
Proprietary
2023/10
59
41
1285
+2/-3
43,788
Meta
Llama 3.1 Community
2023/12
60
36
1284
+2/-2
86,159
Anthropic
Proprietary
2024/4
60
42
1283
+2/-2
63,038
Meta
Llama 3.1 Community
2023/12
61
41
1282
+3/-2
52,144
Proprietary
Online
61
60
1277
+8/-10
4,014
Tencent
Proprietary
N/A
62
79
1282
+2/-3
55,442
xAI
Proprietary
2024/3
62
43
1281
+2/-2
47,973
OpenAI
Proprietary
2023/10
63
63
1279
+3/-4
17,432
Alibaba
Qwen
N/A
63
56
1277
+6/-6
7,451
Meta
Llama
N/A
71
79
1271
+7/-6
7,367
Mistral
Apache 2.0
N/A
72
57
1276
+2/-2
82,435
Proprietary
2023/11
72
74
1274
+3/-3
26,344
DeepSeek
DeepSeek
N/A
72
61
1273
+3/-3
47,631
Meta
Llama-3.3
N/A
72
79
1273
+3/-3
41,519
Alibaba
Qwen
2024/9
73
56
1272
+2/-2
102,133
OpenAI
Proprietary
2023/12
78
86
1260
+10/-10
3,010
Ai2
Llama 3.1
N/A
79
64
1267
+2/-2
48,217
Mistral
Mistral Research
2024/7
79
79
1266
+4/-3
20,580
NexusFlow
CC-BY-NC-4.0
2024/7
79
61
1266
+2/-2
103,748
OpenAI
Proprietary
2023/4
79
79
1265
+3/-3
29,633
Mistral
MRL
N/A
79
62
1258
+9/-8
4,287
Mistral
Proprietary
N/A
80
86
1264
+2/-2
58,637
Meta
Llama 3.1 Community
2023/12
81
58
1263
+2/-1
202,641
Anthropic
Proprietary
2023/8
83
87
1261
+3/-3
26,371
Amazon
Proprietary
N/A
83
65
1261
+2/-2
97,079
OpenAI
Proprietary
2023/12
89
60
1254
+2/-2
49,399
Anthropic
Propretary
N/A
89
86
1251
+6/-7
7,948
Reka AI
Proprietary
N/A
89
88
1246
+7/-10
4,210
Tencent
Proprietary
N/A
92
90
1243
+2/-2
65,661
Proprietary
2023/11
93
88
1237
+4/-6
9,125
AI21 Labs
Jamba Open
2024/3
93
96
1233
+8/-6
5,730
Alibaba
Apache 2.0
N/A
94
89
1236
+2/-2
79,538
Gemma license
2024/6
94
98
1233
+4/-4
15,321
Mistral
Apache 2.0
N/A
94
106
1233
+3/-4
20,646
Amazon
Proprietary
N/A
94
90
1232
+5/-5
10,548
Princeton
MIT
2024/7
94
86
1228
+9/-10
3,889
Nvidia
Llama 3.1
2023/12
95
94
1231
+4/-6
10,535
Cohere
CC-BY-NC-4.0
2024/8
96
110
1228
+3/-3
37,697
Proprietary
N/A
97
106
1222
+9/-11
3,460
Allen AI
Apache-2.0
N/A
99
105
1225
+3/-3
28,768
Cohere
CC-BY-NC-4.0
N/A
99
96
1225
+3/-4
20,608
Nvidia
NVIDIA Open Model
2023/6
99
99
1222
+5/-5
10,221
Zhipu AI
Proprietary
N/A
101
96
1221
+5/-5
8,132
Reka AI
Proprietary
N/A
102
110
1221
+4/-4
25,213
Microsoft
MIT
N/A
103
97
1222
+2/-1
163,629
Meta
Llama 3 Community
2023/12
106
96
1217
+2/-2
113,067
Anthropic
Proprietary
2023/8
109
119
1214
+3/-3
20,654
Amazon
Proprietary
N/A
111
120
1205
+10/-10
2,901
Tencent
Proprietary
N/A
112
121
1201
+10/-9
3,074
Ai2
Llama 3.1
N/A
113
109
1208
+2/-2
57,197
Gemma license
2024/6
113
106
1206
+2/-2
80,846
Cohere
CC-BY-NC-4.0
2024/3
114
109
1203
+3/-2
38,872
Alibaba
Qianwen LICENSE
2024/6
114
92
1202
+2/-3
55,962
OpenAI
Proprietary
2021/9
114
119
1198
+7/-6
5,111
Mistral
MRL
N/A
115
121
1196
+7/-4
10,391
Cohere
CC-BY-NC-4.0
N/A
116
110
1196
+4/-5
10,851
Cohere
CC-BY-NC-4.0
2024/8
117
111
1195
+2/-2
122,309
Anthropic
Proprietary
2023/8
117
105
1194
+4/-4
15,753
DeepSeek AI
DeepSeek License
2024/6
117
120
1192
+5/-5
9,274
AI21 Labs
Jamba Open
2024/3
118
136
1192
+2/-2
52,578
Meta
Llama 3.1 Community
2023/12
126
105
1179
+2/-2
91,614
OpenAI
Proprietary
2021/9
126
121
1177
+3/-3
27,430
Alibaba
Qianwen LICENSE
2024/4
126
153
1169
+11/-10
3,410
Alibaba
Apache 2.0
N/A
127
136
1173
+4/-3
25,135
01 AI
Apache-2.0
2024/5
127
120
1173
+2/-2
64,926
Mistral
Proprietary
N/A
127
121
1172
+4/-5
16,027
Reka AI
Proprietary
Online
130
131
1168
+2/-2
109,056
Meta
Llama 3 Community
2023/3
130
143
1165
+4/-5
10,599
InternLM
Other
2024/8
131
125
1164
+2/-3
56,398
Cohere
CC-BY-NC-4.0
2024/3
131
131
1164
+3/-3
35,556
Mistral
Proprietary
N/A
131
124
1163
+3/-2
53,751
Mistral
Apache 2.0
2024/4
131
127
1163
+4/-4
25,803
Reka AI
Proprietary
2023/11
131
125
1163
+3/-2
40,658
Alibaba
Qianwen LICENSE
2024/2
131
128
1159
+8/-11
3,289
IBM
Apache 2.0
N/A
133
143
1160
+2/-3
48,892
Gemma license
2024/7
140
124
1147
+4/-4
18,800
Proprietary
2023/4
140
134
1143
+8/-7
4,854
HuggingFace
Apache 2.0
2024/4
141
137
1141
+3/-4
22,765
Alibaba
Qianwen LICENSE
2024/2
141
145
1135
+8/-10
3,380
IBM
Apache 2.0
N/A
142
143
1139
+4/-3
26,105
Microsoft
MIT
2023/10
142
154
1135
+3/-5
16,676
Nexusflow
Apache-2.0
2024/3
145
143
1130
+2/-2
76,126
Mistral
Apache 2.0
2023/12
145
149
1127
+4/-6
15,917
01 AI
Yi License
2023/6
145
133
1126
+6/-7
6,557
Proprietary
2023/4
146
147
1125
+4/-4
18,687
Alibaba
Qianwen LICENSE
2024/2
147
147
1122
+6/-7
8,383
Microsoft
Llama 2 Community
2023/8
148
133
1122
+2/-2
68,867
OpenAI
Proprietary
2021/9
148
143
1119
+3/-3
33,743
Databricks
DBRX LICENSE
2023/12
148
151
1119
+7/-6
8,390
Meta
Llama 3.2
2023/12
148
151
1118
+4/-4
18,476
Microsoft
MIT
2023/10
149
152
1115
+6/-6
6,658
AllenAI/UW
AI2 ImpACT Low-risk
2023/11
152
143
1109
+8/-6
7,002
IBM
Apache 2.0
N/A
156
162
1109
+3/-3
39,595
Meta
Llama 2 Community
2023/7
156
149
1107
+4/-5
12,990
OpenChat
Apache-2.0
2024/1
156
156
1107
+4/-4
22,936
LMSYS
Non-commercial
2023/8
157
149
1106
+3/-3
34,173
Snowflake
Apache 2.0
2024/4
157
160
1104
+4/-5
10,415
UC Berkeley
CC-BY-NC-4.0
2023/11
157
166
1100
+7/-9
3,836
NousResearch
Apache-2.0
2024/1
158
151
1100
+4/-4
25,070
Gemma license
2024/2
158
165
1097
+9/-9
3,636
Nvidia
Llama 2 Community
2023/11
162
151
1093
+9/-8
4,988
DeepSeek AI
DeepSeek License
2023/11
162
151
1092
+7/-6
8,106
OpenChat
Apache-2.0
2023/11
163
153
1090
+7/-8
5,088
NousResearch
Apache-2.0
2023/11
163
158
1090
+8/-7
7,191
IBM
Apache 2.0
N/A
164
169
1088
+4/-3
20,067
Mistral
Apache-2.0
2023/12
164
168
1087
+4/-5
12,808
Microsoft
MIT
2023/10
164
169
1086
+9/-7
4,872
Alibaba
Qianwen LICENSE
2024/2
164
165
1078
+14/-14
1,714
Cognitive Computations
Apache-2.0
2023/10
165
145
1083
+4/-4
17,036
OpenAI
Proprietary
2021/9
166
168
1078
+9/-9
4,286
Upstage AI
CC-BY-NC-4.0
2023/11
167
173
1082
+3/-4
21,097
Microsoft
MIT
2023/10
169
174
1079
+4/-4
19,722
Meta
Llama 2 Community
2023/7
172
169
1075
+7/-7
7,176
Microsoft
Llama 2 Community
2023/7
175
179
1070
+7/-6
8,523
Meta
Llama 3.2
2023/12
176
178
1069
+6/-4
11,321
HuggingFace
MIT
2023/10
176
172
1062
+11/-11
2,375
HuggingFace
Apache 2.0
N/A
176
169
1061
+11/-12
2,644
MosaicML
CC-BY-NC-SA-4.0
2023/6
176
177
1057
+15/-15
1,192
Meta
Llama 2 Community
2024/1
177
173
1056
+12/-13
1,811
HuggingFace
MIT
2023/10
180
178
1059
+6/-7
7,509
Meta
Llama 2 Community
2023/7
180
168
1050
+15/-15
1,327
TII
Falcon-180B TII License
2023/9
181
172
1058
+4/-4
19,775
LMSYS
Llama 2 Community
2023/7
181
178
1053
+4/-6
9,176
Gemma license
2024/2
181
178
1053
+5/-5
21,622
Microsoft
MIT
2023/10
181
193
1053
+5/-5
14,532
Meta
Llama 2 Community
2023/7
181
171
1051
+8/-8
5,065
Alibaba
Qianwen LICENSE
2023/8
181
181
1049
+11/-11
2,996
UW
Non-commercial
2023/5
190
182
1037
+6/-5
11,351
Gemma license
2024/2
191
185
1033
+7/-10
5,276
Together AI
Apache 2.0
2023/12
191
199
1031
+8/-7
6,503
Allen AI
Apache-2.0
2024/2
194
192
1023
+5/-5
9,142
Mistral
Apache 2.0
2023/9
194
193
1021
+6/-8
7,017
LMSYS
Llama 2 Community
2023/7
194
182
1019
+7/-5
8,713
Proprietary
2021/6
199
197
1005
+8/-9
4,918
Gemma license
2024/2
199
194
1004
+5/-7
7,816
Alibaba
Qianwen LICENSE
2024/2
201
200
980
+8/-6
7,020
UC Berkeley
Non-commercial
2023/4
201
201
971
+7/-9
4,763
Tsinghua
Apache-2.0
2023/10
202
201
948
+16/-16
1,788
Nomic AI
Non-commercial
2023/3
203
201
944
+9/-9
3,997
MosaicML
CC-BY-NC-SA-4.0
2023/5
203
206
940
+10/-10
2,713
Tsinghua
Apache-2.0
2023/6
203
206
937
+9/-8
4,920
RWKV
Apache 2.0
2023/4
207
201
917
+9/-7
5,864
Stanford
Non-commercial
2023/3
207
207
909
+9/-7
6,368
OpenAssistant
Apache 2.0
2023/4
208
209
895
+8/-10
4,983
Tsinghua
Non-commercial
2023/3
209
209
884
+9/-9
4,288
LMSYS
Apache 2.0
2023/4
211
212
856
+11/-11
3,336
Stability AI
CC-BY-NC-SA-4.0
2023/4
211
209
838
+10/-10
3,480
Databricks
MIT
2023/4
212
210
815
+14/-9
2,446
Meta
Non-commercial
2023/2
Rank (UB): Ranking calculated based on the Bradley-Terry model. This ranking reflects the model's overall performance in the arena and provides an upper bound estimate of its Elo score, helping to understand the model's potential competitiveness.
Rank (StyleCtrl): Ranking after controlling for conversational style. This ranking aims to reduce preference bias caused by the model's response style (e.g., verbosity, conciseness) and more purely assess the model's core capabilities.
Model Name: The name of the Large Language Model (LLM). This column has embedded links to the models; click to navigate.
Score: The Elo rating obtained by the model through user votes in the arena. Elo rating is a relative ranking system, where a higher score indicates better model performance. This score is dynamic, reflecting the model's relative strength in the current competitive environment.
Confidence Interval: The 95% confidence interval for the model's Elo rating (e.g., +6/-6
). A smaller interval indicates more stable and reliable model scores; conversely, a larger interval may indicate insufficient data or greater fluctuation in model performance. It provides a quantitative assessment of the score's accuracy.
Votes: The total number of votes received by the model in the arena. More votes generally mean higher statistical reliability of its score.
Provider: The organization or company providing the model.
License Agreement: The type of license agreement for the model, such as Proprietary, Apache 2.0, MIT, etc.
Knowledge Cutoff Date: The knowledge cutoff date for the model's training data. N/A indicates that the relevant information is not provided or unknown.
This leaderboard data is automatically generated and provided by the fboulnois/llm-leaderboard-csv project, which retrieves and processes data from lmarena.ai. This leaderboard is automatically updated daily by GitHub Actions.
This report is for reference only. Leaderboard data is dynamic and based on user preference votes on Chatbot Arena during a specific period. The completeness and accuracy of the data depend on the upstream data sources and the updates and processing of the fboulnois/llm-leaderboard-csv
project. Different models may use different license agreements; please refer to the official documentation of the model provider when using them.