Skip to content

Configuration of Chat Services

Rodel Agent supports a variety of mainstream AI chat services with different configurations. This document explains how to configure the supported AI services.

Please use the subsection index to navigate to the segment for the service you want to configure.

OpenAI

Documentationhttps://platform.openai.com/docs/api-reference/chat/create
API Tokenhttps://platform.openai.com/account/api-keys

Configuring OpenAI is relatively simple. You just need to input your API Key in the Access Key field of the settings.

Proxy and API Compatibility

Currently, the OpenAI interface data structure has become a de facto standard. Many AI services expose their interfaces by adopting similar address and data structures to OpenAI for easier usage. This is referred to as the OpenAI-compatible interface.

If the AI service you are using is not listed in Rodel Agent's support list but uses an OpenAI-compatible interface, you can input the service's address in the Endpoint (API) text box of the OpenAI settings block.

Another usage scenario involves using OpenAI services through a proxy server due to inaccessibility in some countries or regions, such as OpenAI API Proxy.

In this case, you still need to use your own API key, but instead of directly accessing the OpenAI server, you use a proxy service to avoid restrictions or bans.

You can input the service address in the Endpoint (API) field to achieve API proxy functionality. For example, for the proxy service mentioned above, you need to input the address: https://api.openai-proxy.com/v1.

WARNING

The OpenAI API Proxy mentioned above is provided for illustration purposes only. The developer does not take responsibility for its security. Users need to independently evaluate the reliability of the service and take responsibility for its usage.

TIP

For proxies, including a version number is usually necessary, such as v1 in https://api.openai-proxy.com/v1. Because the API input field needs to be compatible with other API services, some services might not use a version number in the request path.

Azure OpenAI

Documentationhttps://learn.microsoft.com/azure/ai-services/openai/overview
Azure Serviceshttps://azure.microsoft.com/en-us/products/ai-services/openai-service

Microsoft has a special relationship with OpenAI, akin to an authorized dealer, allowing you to create OpenAI resources on the Azure cloud service platform and deploy models such as GPT-3.5 and GPT-4o as needed.

Though their models are consistent, the network request formats and required configuration items differ.

After deploying OpenAI resources, you can find the required keys (either one of the two keys) and endpoint in the resource page under Resource Management -> Keys and Endpoints.

Keys and Endpoints

Input the corresponding values in the application settings.

You might be curious about the API version; here is the documentation. Generally, it doesn't need to be changed.

Adding Custom Models

Compared to OpenAI, Azure OpenAI requires manual deployment of the needed models.

In the application, Azure OpenAI does not provide pre-configured models. Simply filling in the keys and endpoints will not make the configuration effective; we also need to create custom models.

In the Azure OpenAI Studio, you can deploy specific models from the model library.

Each model has its own Id, for instance, the Id for GPT-3.5 Turbo is gpt-35-turbo.

Azure Model Deployment

When deploying a model, Azure requires you to provide a Deployment Name, which is crucial. When interacting with the service through the API, the model identifier is not the model ID but the deployment name you provided during the model deployment.

It is recommended to use the same name as the model ID during deployment to avoid confusion. However, exceptions may occur, such as when you deploy multiple GPT-3.5 Turbo models with different versions. In that case, your deployment names might be gpt-35-turbo-1106 or gpt-35-turbo-0613. When creating a custom model in the application, you can do it as shown below:

Custom Chat Model

Azure AI Foundry

Azure offers a comprehensive AI model distribution service—Azure AI Foundry.

Assuming you have already created your project in Azure AI Foundry, click on My assets / Models + endpoints in the sidebar.

Manage model and service deployments

It is important to note that the models deployed in Azure AI Foundry are categorized, and you can see two types in the image above.

When filling in the application configuration, the model and category must match.

For example, if we want to add the Phi-4 model to the application, follow these steps:

  1. Click the Get endpoint button in the category to which Phi-4 belongs in the image above.
    Model Endpoint
  2. Copy the endpoint and key to the corresponding input box in the application.
  3. Create a custom model, where the model ID corresponds to your deployment name (the Name in the first column).

Gemini

Documentationhttps://ai.google.dev/gemini-api/docs
API Tokenhttps://ai.google.dev/gemini-api/docs/api-key

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

Anthropic

Documentationhttps://docs.anthropic.com/zh-CN/docs/intro-to-claude
API Tokenhttps://console.anthropic.com/account/keys

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

Moonshot

Documentationhttps://platform.moonshot.cn/docs/api/chat#api-说明
API Tokenhttps://platform.moonshot.cn/console/api-keys

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

Zhipu AI

ChatGLM is from here.

Documentationhttps://open.bigmodel.cn/dev/howuse/introduction
API Tokenhttps://open.bigmodel.cn/usercenter/apikeys

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

01.AI

Documentationhttps://platform.lingyiwanwu.com/docs
API Tokenhttps://platform.lingyiwanwu.com/apikeys

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

DeepSeek

Documentationhttps://platform.deepseek.com/api-docs/zh-cn
API Tokenhttps://platform.deepseek.com/api_keys

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

Qwen

Documentationhttps://help.aliyun.com/zh/model-studio/developer-reference/use-qwen-by-calling-api
API Tokenhttps://help.aliyun.com/zh/model-studio/developer-reference/get-api-key

Qwen is a publicly available large language model by Alibaba, hosted on Alibaba Cloud's large model service platform, Bailian. Therefore, you need to register and activate the model service on the Alibaba Cloud Bailian platform.

The Bailian platform has its own API interface and data structure. However, a notable feature is that it also provides an Open AI compatible interface. Rodel Agent utilizes the compatible interface provided by Bailian, specifically:

https://dashscope.aliyuncs.com/compatible-mode/v1

Thus, the supported models are limited. For specific information, refer to this documentation: OpenAI Interface Compatibility

ERNIE Bot

Documenthttps://cloud.baidu.com/doc/WENXINWORKSHOP/s/fm4tsw9nv
API Tokenhttps://console.bce.baidu.com/iam/#/iam/accesslist

ERNIE Bot is a large language model provided by Baidu, hosted on Baidu Cloud's Qianfan Large Model Platform. It has its own hosting steps, and services need to be activated separately for different models. Please refer to its documentation for detailed steps.

TIP

Personally, it's hard for me to evaluate Baidu Qianfan's documentation; I can only say it's the most misleading among these 20+ services. During the integration of ERNIE Bot, I often felt like giving up. So if you want to save some trouble, just click the API Token link and create an Access Key. Fill the Access Key and Secret Key into the application's Access Key and Secret Key fields respectively.

Ignore the API documentation's mentions of API Key, Register Application, and other miscellaneous things.

Tencent Hunyuan

Documentationhttps://cloud.tencent.com/document/product/1729/105701
API Tokenhttps://cloud.tencent.com/document/product/1729/111008

The Hunyuan large model is hosted on Tencent Cloud. A notable advantage is that it provides Open AI compatible interfaces, allowing direct integration with Rodel Agent.

You can create an API Key, then fill in the key into the access credentials of the Rodel Agent.

iFLYTEK Spark

Documentationhttps://www.xfyun.cn/doc/spark/Web.html
API Tokenhttps://console.xfyun.cn/services/bm35

The Spark large model is a large language model launched by iFLYTEK. Before using this model family, you need to register an application on the iFLYTEK Open Platform and then activate the corresponding model service.

Models that are not activated cannot be used.

After registration, you can find the Service Interface Authentication Information on the model page of Spark Cognitive Large Model.

Rodel Agent uses the http service. You only need to fill in the APIPassword of the corresponding model into Rodel Agent's Access Key.

ByteDance Doubao

Documentationhttps://www.volcengine.com/docs/82379/1263482
API Tokenhttps://console.volcengine.com/ark/region:ark+cn-beijing/apiKey

The Doubao model is a large language model launched by ByteDance. Before using this model, you need to register an account on VolcEngine and then create a reasoning access point on the model reasoning page on Volcano方舟.

You need to manually deploy the model access point, and VolcEngine does not provide a preset reasoning access point.

After that, you need to create an API Key in API Key Management, and fill it in the settings of the application.

Custom Models

Please refer to Create a Reasoning Access Point to deploy the model first.

Then click Create Custom Model in the Doubao setting section, paste the reasoning access point ID (ep-xxxxxxx) into the model ID, and give a descriptive name to the custom model. In this way, a custom model is created, and you can start chatting with this model.

Silicon Flow

Documentationhttps://openrouter.ai/docs/quick-start
API Tokenhttps://docs.siliconflow.cn/docs/4-api调用

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

OpenRouter

Documentationhttps://openrouter.ai/docs/quick-start
API Tokenhttps://openrouter.ai/keys

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

Together AI

Documentationhttps://docs.together.ai/docs/quickstart
API Tokenhttps://api.together.xyz/settings/api-keys

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

Groq

Documentationhttps://console.groq.com/docs/quickstart
API Tokenhttps://console.groq.com/keys

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

Perplexity

Documentationhttps://docs.perplexity.ai/docs/getting-started
API Tokenhttps://www.perplexity.ai/settings/api

Similar to OpenAI, after obtaining the access key, fill it in the Access Key field.

Mistral AI

Documenthttps://docs.mistral.ai/
API Tokenhttps://console.mistral.ai/api-keys/
Codestral Tokenhttps://console.mistral.ai/codestral

Similar to Open AI, after obtaining the access key, enter it into the Access Key field.

Mistral also offers the Codestral service, which is free (at least until 2025/03/22) and primarily used for code generation (though it can also handle conversations).

Since they use different APIs, Rodel Agent provides a separate key location for Codestral. If you need to use Codestral, enter the Codestral key and then turn on the Use Codestral switch. When calling the model, choose Codestral.

Ollama

Documentationhttps://github.com/ollama/ollama/blob/main/README.md#quickstart
Service AddressDefault is http://localhost:11434/v1
API TokenDefault is ollama

Ollama is a popular local model hosting service. Its usage itself is extensive enough to warrant a lengthy document, which is not covered here. Please refer to their documentation.

Rodel Agent supports accessing Ollama via API. Ollama provides an OpenAI-compatible interface, allowing simple integration with Rodel Agent.

Since it runs locally, the application does not include any initial models. You need to create custom models based on the models already pulled to your local environment.

Creating Custom Models

Assuming you just installed Ollama and plan to run the Qwen 2 (Qwen 2) large model, follow these steps:

  1. Pull the model to your local environment using Ollama (using qwen2 0.5b as an example)

    shell
    ollama pull qwen2:0.5b
  2. Confirm the model has been downloaded locally by executing:

    shell
    ollama list

    It should return content similar to the following:

    shell
    NAME                    ID              SIZE    MODIFIED
    qwen2:0.5b              6f48b936a09f    352 MB  36 seconds ago
  3. Create a custom model in the application based on the model name:

    Custom Chat Model

Released under the GPLv3 License