Back to Templates
  • +6

OpenAI Responses API Adapter for LLM and AI Agent Workflows

Created by

Jimleuk

Last update

Last update 15 days ago

Categories
Share

This n8n template demonstrates how to use OpenAI's Responses API with existing LLM and AI Agent nodes.

Though I would recommend just waiting for official support, if you're impatient and would like a round-about way to integrate OpenAI's responses API into your existing AI workflows then this template is sure to satisfy!

This approach implements a simple API wrapper for the Responses API using n8n's builtin webhooks. When the base url is pointed to these webhooks using a custom OpenAI credential, it's possible to intercept the request and remap for compatibility.

How it works

  • An OpenAI subnode is attached to our agent but has a special custom credential where the base_url is changed to point at this template's webhooks.
  • When executing a query, the agent's request is forwarded to our mini chat completion workflow.
  • Here, we take the default request and remap the values to use with a HTTP node which is set to query the Responses API.
  • Once a response is received, we'll need to remap the output for Langchain compatibility. This just means the LLM or Agent node can parse it and respond to the user.
  • There are two response formats, one for streaming and one for non-streaming responses.

How to use

  • You must activate this workflow to be able to use the webhooks.
  • Create the custom OpenAI credential as instructed.
  • Go to your existing AI workflows and replace the LLM node with the custom OpenAI credential. You do not need to copy anything else over to the existing template.

Requirements

  • OpenAI account for Responses API

Customising this workflow

  • Feel free to experiment with other LLMs using this same technique!
  • Keep up to date with the Responses API announcements and make modifications as required.