OpenAI
Call FastMCP servers from the OpenAI API
OpenAI recently announced support for MCP servers in the Responses API. Note that at this time, MCP is not supported in ChatGPT.
Responses API
OpenAI’s Responses API supports MCP servers as remote tool sources, allowing you to extend AI capabilities with custom functions.
The Responses API is a distinct API from OpenAI’s Completions API, Assistants API, or ChatGPT. At this time, only the Responses API supports MCP.
Currently, the Responses API only accesses tools from MCP servers—it queries the list_tools
endpoint and exposes those functions to the AI agent. Other MCP features like resources and prompts are not currently supported.
Create a Server
First, create a FastMCP server with the tools you want to expose. For this example, we’ll create a server with a single tool that rolls dice.
Deploy the Server
Your server must be deployed to a public URL in order for OpenAI to access it.
For development, you can use tools like ngrok
to temporarily expose a locally-running server to the internet. We’ll do that for this example (you may need to install ngrok
and create a free account), but you can use any other method to deploy your server.
Assuming you saved the above code as server.py
, you can run the following two commands in two separate terminals to deploy your server and expose it to the internet:
This exposes your unauthenticated server to the internet. Only run this command in a safe environment if you understand the risks.
Call the Server
To use the Responses API, you’ll need to install the OpenAI Python SDK (not included with FastMCP):
You’ll also need to authenticate with OpenAI. You can do this by setting the OPENAI_API_KEY
environment variable. Consult the OpenAI SDK documentation for more information.
Here is an example of how to call your server from Python. Note that you’ll need to replace https://your-server-url.com
with the actual URL of your server. In addition, we use /sse
as the endpoint because we deployed an SSE server with the default path; you may need to use a different endpoint if you customized your server’s deployment.
If you run this code, you’ll see something like the following output:
Authentication
New in version: 2.6.0
The Responses API can include headers to authenticate the request, which means you don’t have to worry about your server being publicly accessible.
Server Authentication
The simplest way to add authentication to the server is to use a bearer token scheme.
For this example, we’ll quickly generate our own tokens with FastMCP’s RSAKeyPair
utility, but this may not be appropriate for production use. For more details, see the complete server-side Bearer Auth documentation.
We’ll start by creating an RSA key pair to sign and verify tokens.
FastMCP’s RSAKeyPair
utility is for development and testing only.
Next, we’ll create a BearerAuthProvider
to authenticate the server.
Here is a complete example that you can copy/paste. For simplicity and the purposes of this example only, it will print the token to the console. Do NOT do this in production!
Client Authentication
If you try to call the authenticated server with the same OpenAI code we wrote earlier, you’ll get an error like this:
As expected, the server is rejecting the request because it’s not authenticated.
To authenticate the client, you can pass the token in the Authorization
header with the Bearer
scheme:
You should now see the dice roll results in the output.