The 30-Second Version (read this first)
- Any code, script, desktop app or SaaS that already talks to the OpenAI chat.completions endpoint will work with CustomGPT.ai RAG API.
- The “porting” is literally two lines of code or two environment variables.
Read the OpenAI compatibility docs here.
from openai import OpenAI # keep using the official SDK
client = OpenAI(
api_key="CUSTOMGPT_API_KEY", # ← your key
base_url="https://app.customgpt.ai/api/v1/projects/{project_id}/" # ← your project
)
Or (for most CLIs / Docker images):
export OPENAI_API_KEY="CUSTOMGPT_API_KEY"
export OPENAI_API_BASE="https://app.customgpt.ai/api/v1/projects/{project_id}/"
- Result: every request now goes to CustomGPT.ai’s Retrieval-Augmented Generation (RAG) backend, so the same tools suddenly answer with grounded, non-hallucinating responses from your documents instead of a generic model.
List TLDR:
- SDKs & Core Frameworks – LlamaIndex, Langchain, etc
- Visual & Low-Code Builders – Flowise, Langflow, etc
- Autonomous & Multi-Agent Frameworks – Auto-GPT, BabyAGI, etc
- IDE & Developer Productivity – Continue, CodeGPT, etc
- Voice & Audio Pipelines – Pipecat, Whisper, telegram bots, etc
- Automation / iPaaS
- OpenAI-Compatible Model Servers
- Knowledge, Notes & RAG Front-Ends
- Messaging-App Bots
- Browser & OS Helpers
- Dashboards & Observability
Compatibility details & limitations: only chat.completions is supported; model name and most tuning params are ignored.
Why this matters
- Zero-friction upgrade. No need to rewrite prompts, chain logic, or pipeline code you already maintain.
- Instant RAG. Tools that never had retrieval—IDEs, browser extensions, voice assistants—now inherit CustomGPT.ai’s accuracy and citation support.
- Same ecosystem. Continue using LangChain, LlamaIndex, vLLM, Zapier, VS Code extensions, etc. They still think they are talking to OpenAI.
Quick Reference: What Just Works
Below is a condensed catalogue of the 100 + open-source and commercial projects you can repoint today. They are grouped so you can scan for the category you care about; the GitHub link (or vendor URL) is included for each.
1. SDKs & Core Frameworks
These are the basic tools developers use to build AI applications. Adding RAG helps these tools give answers based on your own documents instead of making things up.
This means you get responses that reference your specific information rather than generic answers.
| Tool | What it is | GitHub / Docs |
| OpenAI Python SDK | Official client | https://github.com/openai/openai-python |
| OpenAI Node / TS SDK | JavaScript client | https://github.com/openai/openai-node |
| OpenAI .NET SDK | C# / F# client | https://github.com/openai/openai-dotnet |
| OpenAI Java SDK | Java / Kotlin client | https://github.com/openai/openai-java |
| Go-OpenAI | Popular Go wrapper | https://github.com/sashabaranov/go-openai |
| openai-php/client | PHP SDK | https://github.com/openai-php/client |
| ruby-openai | Ruby gem | https://github.com/alexrudall/ruby-openai |
| OpenAI-Swift | iOS / macOS SDK | https://github.com/MacPaw/OpenAI |
| LangChain | Prompt / agent framework | https://github.com/langchain-ai/langchain |
| LlamaIndex | Data-centric RAG | https://github.com/run-llama/llama_index |
| Haystack | Production RAG pipelines | https://github.com/deepset-ai/haystack |
| Semantic-Kernel | .NET / Python orchestration | https://github.com/microsoft/semantic-kernel |
| CopilotKit | React AI copilots | https://github.com/CopilotKit/CopilotKit |
2. Visual & Low-Code Builders
These easy-to-use tools let people without coding skills build AI applications.
With RAG, these tools can now look up information in your documents to give accurate answers based on your specific content, not just general knowledge.
| Tool | Purpose | GitHub / Website |
| Flowise | Drag-and-drop LangChain flows | https://github.com/FlowiseAI/Flowise |
| LangFlow | Visual agent designer | https://github.com/logspace-ai/langflow |
| Chainlit | One-file chat UI | https://github.com/Chainlit/chainlit |
| Dify | End-to-end LLM platform | https://github.com/langgenius/dify |
| Open WebUI | Self-hosted ChatGPT panel | https://github.com/open-webui/open-webui |
| ChatGPT-Next-Web | Vercel-ready chat site | https://github.com/Yidadaa/ChatGPT-Next-Web |
| ChatBox | Electron desktop UI | https://github.com/ChatBoxAI/ChatBox |
| Lencx ChatGPT | Native (Tauri) client | https://github.com/lencx/ChatGPT |
| Chuanhu ChatGPT | PySide desktop UI | https://github.com/GaiZhenbiao/ChuanhuChatGPT |
| ChatALL | Multi-model comparer | https://github.com/sunner/ChatALL |
3. Autonomous & Multi-Agent Frameworks
These are AI systems that work on their own or in teams. By connecting them to RAG, they can search through your documents to find specific information.
This makes them much smarter about your particular business instead of relying on general knowledge.
4. IDE & Developer Productivity
These tools help programmers write code faster. Adding RAG means they can now suggest code based on your own documentation and coding standards.
This helps developers follow your team’s best practices instead of using generic code examples.
5. Voice & Audio Pipelines
These tools convert speech to text and let you talk with AI. With RAG, these voice assistants can answer questions using information from your documents, giving you accurate answers about your specific business when you ask questions.
| Tool | Function | GitHub |
| Pipecat | Real-time voice agent | https://github.com/pipecat-ai/pipecat |
| Whisper (official) | Speech-to-text model | https://github.com/openai/whisper |
| whisper.cpp | Fast local STT | https://github.com/ggerganov/whisper.cpp |
| whisper-cli | STT command-line | https://github.com/vatsalaggarwal/whisper-cli |
| MeetingSummarizer | Record → GPT summary | https://github.com/rajpdus/MeetingSummarizer |
| OpenAI Realtime Twilio | Phone ↔ GPT demo | https://github.com/openai/openai-realtime-twilio-demo |
| speech-to-speech (HF) | Full voice pipeline | https://github.com/huggingface/speech-to-speech |
| Telegram voice bots (× 4) | Voice-note → text | eg. https://github.com/gsamat/telegram_openai_whisper_api_transcriber |
6. Automation / iPaaS
These tools connect different systems and automate tasks. With RAG, they can make better decisions based on your company’s own rules and documents. This means automated workflows will follow your specific business processes correctly.
| Tool | Nature | Link |
| n8n | Self-hosted Zapier alt | https://github.com/n8n-io/n8n |
| Activepieces | MIT no-code flows | https://github.com/activepieces/activepieces |
| Node-RED | Flow-based wiring | https://nodered.org |
| Zapier NLA | Natural-language actions | https://nla.zapier.com |
| Huginn | Self-hosted IFTTT | https://github.com/huginn/huginn |
| LiteLLM | Multi-LLM proxy | https://github.com/BerriAI/litellm |
| Portkey | Reliability gateway | https://github.com/portkey-ai/portkey-node |
7. OpenAI-Compatible Model Servers
These are servers that run AI models similar to OpenAI’s. By connecting them to RAG, you can keep using your existing setup while adding the ability to search through your documents for better answers.
| Tool | Use-case | GitHub |
| vLLM | High-throughput GPU server | https://github.com/vllm-project/vllm |
| LocalAI | Run GGUF/GGML models | https://github.com/go-skynet/LocalAI |
| OpenLLM | BentoML one-command | https://github.com/bentoml/OpenLLM |
| FastChat | Vicuna & proxy mode | https://github.com/lmsysorg/FastChat |
| Oobabooga Text-Gen UI | Web UI + OpenAI API ext. | https://github.com/oobabooga/text-generation-webui |
8. Knowledge, Notes & RAG Front-Ends
These tools are already designed for organizing and finding information. RAG makes them even better by improving search results and adding the ability to show exactly where information came from in your documents.
| Tool | Focus | GitHub |
| DocsGPT | Ask docs repos | https://github.com/arc53/DocsGPT |
| private-gpt | Local RAG API | https://github.com/zylon-ai/private-gpt |
| Quivr | “Second brain” | https://github.com/QuivrHQ/quivr |
| Obsidian Copilot | Vault chat | https://github.com/millsms/obsidian-copilot |
| pdfGPT | Chat with PDFs | https://github.com/bhaskatripathi/pdfGPT |
| GPT-Researcher | Autonomous research | https://github.com/assafelovic/gpt-researcher |
| Rasa | Conversational AI | https://github.com/RasaHQ/rasa |
| RasaGPT | Rasa + LangChain boilerplate | https://github.com/paulpierre/RasaGPT |
9. Messaging-App Bots
These AI assistants work in chat apps like Slack or Discord. With RAG, these bots can answer questions using information from your company documents, helping team members get accurate information without leaving their chat app.
| Platform | Example repo | Link |
| Discord | GPT Discord Bot | https://github.com/openai/gpt-discord-bot |
| Discord | chatgpt-discord | https://github.com/0xk1h0/chatgpt-discord |
| Telegram | chatgpt_telegram_bot | https://github.com/karfly/chatgpt_telegram_bot |
| Telegram | n3d1117 bot | https://github.com/n3d1117/chatgpt-telegram-bot |
| Slack | openai-slack-bot | https://github.com/jack482653/openai-slack-bot |
| Slack | chatgpt-slack-bot | https://github.com/Zeta36/chatgpt-slack-bot |
| whatsapp-gpt | https://github.com/danielgross/whatsapp-gpt | |
| chatgpt-on-wechat | https://github.com/zhayujie/chatgpt-on-wechat |
10. Browser & OS Helpers
These tools add AI features to your web browser or computer. Adding RAG turns them into specialists that understand your specific content, giving you helpful information from your own documents while you browse or work.
| Tool | Quick description | GitHub |
| ChatGPT Raycast | System-wide popup | https://github.com/abielzulio/chatgpt-raycast |
| ChatFred (Alfred) | macOS workflow | https://github.com/chrislemke/ChatFred |
| OpenAI-Translator | Inline translator | https://github.com/openai-translator/openai-translator |
| screenshot-gpt | Explain screen region | https://github.com/PouyaMT/screenshot-gpt |
| screenshot-to-code | UI → Tailwind code | https://github.com/abi/screenshot-to-code |
| qrGPT | Artistic QR codes | https://github.com/Nutlope/qrGPT |
| GPT for Sheets/Docs | Google Workspace add-on | https://workspace.google.com/marketplace/app/gpt_for_sheets_and_docs/677318054654 |
| dallecli | DALL-E in terminal | https://github.com/raiyanyahya/dallecli |
| Email summarizer | Gmail TL;DR | https://github.com/petermartens98/OpenAI-LangChain-Email-Summarizer |
11. Dashboards & Observability
These tools track how well your AI systems are working. With RAG, you can see which documents are being used most often and how people are interacting with your information, helping you improve your knowledge base.
| Tool | Purpose | Link |
| Helicone | Logging / cost tracking proxy | https://github.com/helix-onprem/hcone |
| LangSmith | Prompt / chain debugger | https://smith.langchain.com |
| PromptFlow | Experiment hub (Azure) | https://github.com/microsoft/promptflow |
| gpt-review | GitHub PR reviewer | https://github.com/microsoft/gpt-review |
| code-review-gpt | CLI PR summaries | https://github.com/mattzcarey/code-review-gpt |
How to use the tables
- Find your favourite tool in the category tables.
- Follow its link; open its README or settings panel.
- Replace the OpenAI key & URL with the CustomGPT.ai pair shown at the top.
Every request now hits your CustomGPT.ai project and returns retrieval-augmented, citation-rich answers grounded in the data you uploaded—no further code changes required.
How to Switch Each Tool
Most projects already look for one of these variables:
| Common var | What to put |
| OPENAI_API_KEY | your CustomGPT_API_KEY |
| OPENAI_API_BASE, OPENAI_API_BASE_URL, OPENAI_HOST, OPENAI_BASE_URL | https://app.customgpt.ai/api/v1/projects/{project_id}/ |
If the project uses the official SDK programmatically, pass base_url= when you instantiate the OpenAI client (see snippet above).
Project ID is the your agents id (the one you created using customgpt platform)
What to Expect
- Endpoints: only /v1/chat/completions is exposed today. Embeddings, vision, audio, tools, etc. return 404/501.
- Model field ignored: keep a placeholder (e.g. “gpt-4”); CustomGPT.ai routes to the model tied to your project.
- Parameters like temperature, stop, max_tokens are ignored for now.
- Self-managed history: include the full messages array every call.
- Usage metrics: token counts aren’t returned yet.
Full compatibility matrix is in the docs.
Beyond the Quick Adapter
The two-line shim is ideal for proofs-of-concept or extending existing tools. When you need:
- native conversation state
- batch document ingestion
- advanced RAG controls (chunking, citations, analytics)
switch to the official CustomGPT.ai SDK instead; it exposes the full agent/conversation API set.
Final Thoughts
You already have the tools you love—LangChain chains, VS Code copilot, Raycast extension, voice bots, Zapier zaps. Point them at CustomGPT.ai and you immediately gain enterprise-grade RAG grounded in your private content, with zero vendor lock-in and near-zero code changes.
Happy porting! Let us know in the Slack community if you hit an edge-case—we’re iterating fast.
Frequently Asked Questions
How do I connect existing OpenAI-compatible tools to the CustomGPT RAG API?
You can keep your existing OpenAI SDK code and change only two settings: `OPENAI_API_KEY=` and `OPENAI_API_BASE=https://app.customgpt.ai/api/v1/projects/{project_id}/`. Then send the same Chat Completions call to `https://app.customgpt.ai/api/v1/projects/{project_id}/chat/completions` with headers `Authorization: Bearer ` and `Content-Type: application/json`, plus a minimal body like `{“model”:”gpt-4o-mini”,”messages”:[{“role”:”user”,”content”:”Summarize our return policy”}]}`. What stays unchanged: your OpenAI client methods, message format, and response parsing flow. What changes: key, base URL, and project-specific endpoint. For exact auth and endpoint rules, check your dashboard API Reference under Authentication and RAG endpoints. If `project_id` is missing or invalid, expect a 404 Not Found or project access error. From API usage patterns, temperature 0 to 0.2 improves citation consistency versus Pinecone Assistant and Azure OpenAI setups.
Which types of tools work with the CustomGPT RAG API?
You can treat the live Integrations Catalog as the source of truth for compatibility: as of March 2026, it lists 100+ connectors across 11 categories. You can connect SDKs and frameworks such as LlamaIndex and LangChain, visual builders like Flowise and Dify, plus voice, automation, messaging, and dashboard tools.
If your app already uses OpenAI-style chat completions, you can keep the same request shape and change only the base URL and API key per the OpenAI-compatible API docs, which is the common migration path from OpenAI or Anthropic clients.
Before production rollout, check your exact connector on the pricing page, because some integrations are plan-gated. In sales call transcript analysis, the most frequent launch delay came from connector-plan mismatches, not endpoint changes, so confirming plan access early usually prevents rework.
What are the compatibility limitations of the CustomGPT RAG API?
The CustomGPT RAG API supports only the /v1/chat/completions endpoint. The model name field is ignored (CustomGPT routes to the model tied to your project). Parameters like temperature, stop, and max_tokens are currently ignored. Embeddings, vision, audio, and tools endpoints return 404/501. You must manage conversation history yourself by including the full messages array with each call. Token usage metrics are not returned yet.
Why should I use the CustomGPT RAG API instead of raw OpenAI?
You should choose the CustomGPT RAG API when you need OpenAI-style integration plus private-document grounding without rebuilding prompts, orchestration, or tooling integrations. If you already call OpenAI chat completions, you only change the base URL and API key while keeping your existing message payload shape, roles, and tool-calling schema. You can keep LangChain or LlamaIndex flows as-is, unlike a DIY OpenAI plus Pinecone stack that often needs extra retrieval glue code. Responses are grounded in your indexed internal documents and return source citations, so each claim can be traced to a specific file or passage for audit, compliance, or support review. Based on product benchmark data from enterprise deployments, teams migrating this way reached a first production RAG release in days, not weeks, and most did it without prompt rewrites.
How do I switch an existing OpenAI SDK integration to CustomGPT?
You can migrate by only changing endpoint and key. Example pattern:
`POST https://app.customgpt.ai/api/v1/projects/{project_id}/chat/completions`
Headers: `Authorization: Bearer `, `Content-Type: application/json`
Body: `{“model”:”gpt-4o-mini”,”messages”:[{“role”:”user”,”content”:”Summarize this file”}],”temperature”:0.2}`
If you use env vars, set `OPENAI_API_KEY=` and base URL with project path. SDK names vary by version: `OPENAI_API_BASE`, `OPENAI_BASE_URL`, `OPENAI_API_BASE_URL`, and sometimes `OPENAI_HOST` are recognized. If your client ignores `OPENAI_API_BASE`, pass `base_url` directly in code when creating the client.
Quick checks: 401 usually means wrong key; 404 usually means wrong `{project_id}` or missing `/projects/{project_id}/`. Copy project ID from your agent settings page. A documentation audit and competitive research (against LangChain and Azure OpenAI adapters) also show keeping the trailing slash reduces routing errors in some clients.
Priyansh is Developer Relations Advocate who loves technology, writer about them, creates deeply researched content about them.