mirror of
https://github.com/infiniflow/ragflow.git
synced 2026-03-28 01:50:38 +08:00
### What problem does this PR solve? Adds Perplexity contextualized embeddings API as a new model provider, as requested in #13610. - `PerplexityEmbed` provider in `rag/llm/embedding_model.py` supporting both standard (`/v1/embeddings`) and contextualized (`/v1/contextualizedembeddings`) endpoints - All 4 Perplexity embedding models registered in `conf/llm_factories.json`: `pplx-embed-v1-0.6b`, `pplx-embed-v1-4b`, `pplx-embed-context-v1-0.6b`, `pplx-embed-context-v1-4b` - Frontend entries (enum, icon mapping, API key URL) in `web/src/constants/llm.ts` - Updated `docs/guides/models/supported_models.mdx` - 22 unit tests in `test/unit_test/rag/llm/test_perplexity_embed.py` Perplexity's API returns `base64_int8` encoded embeddings (not OpenAI-compatible), so this uses a custom `requests`-based implementation. Contextualized vs standard model is auto-detected from the model name. Closes #13610 ### Type of change - [x] New Feature (non-breaking change which adds functionality) - [x] Documentation Update
97 lines
10 KiB
Plaintext
97 lines
10 KiB
Plaintext
---
|
|
sidebar_position: 3
|
|
slug: /supported_models
|
|
sidebar_custom_props: {
|
|
categoryIcon: LucideBox
|
|
}
|
|
---
|
|
# Supported models
|
|
|
|
import APITable from '@site/src/components/APITable';
|
|
|
|
A complete list of models supported by RAGFlow, which will continue to expand.
|
|
|
|
```mdx-code-block
|
|
<APITable>
|
|
```
|
|
|
|
| Provider | LLM | Image2Text | Speech2text | TTS | Embedding | Rerank | OCR |
|
|
| --------------------- | ------------------ | ------------------ | ------------------ | ------------------ | ------------------ | ------------------ | ------------------ |
|
|
| Anthropic | :heavy_check_mark: | | | | | | |
|
|
| Avian | :heavy_check_mark: | | | | | | |
|
|
| Azure-OpenAI | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | | :heavy_check_mark: | | |
|
|
| BaiChuan | :heavy_check_mark: | | | | :heavy_check_mark: | | |
|
|
| BaiduYiyan | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| Bedrock | :heavy_check_mark: | | | | :heavy_check_mark: | | |
|
|
| Cohere | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| DeepSeek | :heavy_check_mark: | | | | | | |
|
|
| Fish Audio | | | | :heavy_check_mark: | | | |
|
|
| Gemini | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | | |
|
|
| Google Cloud | :heavy_check_mark: | | | | | | |
|
|
| GPUStack | :heavy_check_mark: | | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| Groq | :heavy_check_mark: | | | | | | |
|
|
| HuggingFace | :heavy_check_mark: | | | | :heavy_check_mark: | | |
|
|
| Jina | | | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| LocalAI | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | | |
|
|
| LongCat | :heavy_check_mark: | | | | | | |
|
|
| LM-Studio | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | | |
|
|
| MiniMax | :heavy_check_mark: | | | | | | |
|
|
| MinerU | | | | | | | :heavy_check_mark: |
|
|
| Mistral | :heavy_check_mark: | | | | :heavy_check_mark: | | |
|
|
| ModelScope | :heavy_check_mark: | | | | | | |
|
|
| Moonshot | :heavy_check_mark: | :heavy_check_mark: | | | | | |
|
|
| NovitaAI | :heavy_check_mark: | | | | :heavy_check_mark: | | |
|
|
| NVIDIA | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| Ollama | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | | |
|
|
| OpenAI | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | | |
|
|
| OpenAI-API-Compatible | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| OpenRouter | :heavy_check_mark: | :heavy_check_mark: | | | | | |
|
|
| Perplexity | | :heavy_check_mark: | | | | | |
|
|
| Replicate | :heavy_check_mark: | | | | :heavy_check_mark: | | |
|
|
| PPIO | :heavy_check_mark: | | | | | | |
|
|
| SILICONFLOW | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| StepFun | :heavy_check_mark: | | | | | | |
|
|
| Tencent Hunyuan | :heavy_check_mark: | | | | | | |
|
|
| Tencent Cloud | | | :heavy_check_mark: | | | | |
|
|
| TogetherAI | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| TokenPony | :heavy_check_mark: | | | | | | |
|
|
| Tongyi-Qianwen | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| Upstage | :heavy_check_mark: | | | | :heavy_check_mark: | | |
|
|
| VLLM | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| VolcEngine | :heavy_check_mark: | | | | | | |
|
|
| Voyage AI | | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| Xinference | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| XunFei Spark | :heavy_check_mark: | | | :heavy_check_mark: | | | |
|
|
| xAI | :heavy_check_mark: | :heavy_check_mark: | | | | | |
|
|
| ZHIPU-AI | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | | |
|
|
| DeepInfra | :heavy_check_mark: | | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | | |
|
|
| 302.AI | :heavy_check_mark: | :heavy_check_mark: | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
| CometAPI | :heavy_check_mark: | | | | :heavy_check_mark: | | |
|
|
| DeerAPI | :heavy_check_mark: | :heavy_check_mark: | | :heavy_check_mark: | :heavy_check_mark: | | |
|
|
| Jiekou.AI | :heavy_check_mark: | | | | :heavy_check_mark: | :heavy_check_mark: | |
|
|
|
|
```mdx-code-block
|
|
</APITable>
|
|
```
|
|
|
|
:::danger IMPORTANT
|
|
If your model is not listed here but has APIs compatible with those of OpenAI, click **OpenAI-API-Compatible** on the **Model providers** page to configure your model.
|
|
:::
|
|
|
|
## Example: AI Badgr (OpenAI-compatible)
|
|
|
|
You can use **AI Badgr** with RAGFlow via the existing OpenAI-API-Compatible provider.
|
|
|
|
To configure AI Badgr:
|
|
|
|
- **Provider**: `OpenAI-API-Compatible`
|
|
- **Base URL**: `https://aibadgr.com/api/v1`
|
|
- **API Key**: your AI Badgr API key (from the AI Badgr dashboard)
|
|
- **Model**: any AI Badgr chat or embedding model ID, as exposed by AI Badgr's OpenAI-compatible APIs
|
|
|
|
AI Badgr implements OpenAI-compatible endpoints for `/v1/chat/completions`, `/v1/embeddings`, and `/v1/models`, so no additional code changes in RAGFlow are required.
|
|
|
|
:::note
|
|
The list of supported models is extracted from [this source](https://github.com/infiniflow/ragflow/blob/main/rag/llm/__init__.py) and may not be the most current. For the latest supported model list, please refer to the Python file.
|
|
:::
|