Skip to content

hekmon/comfyui-openai-api

Repository files navigation

ComfyUI OpenAI API

This repository contains ComfyUI nodes that integrates with the OpenAI API: it allows you to use language models and vision language models within your workflow.

It is KISS by design and intended for those who only wants basic capabilities without having to import massive projects like LLM party: only the chat completions endpoint (with vision support) is implemented as it should be enough for 99% of use cases.

Thanks to its simplicity the project has a low footprint: it only has 1 external dependency (3 in total) !

  • openai the official OpenAI API bindings for Python
  • numpy for computation but it is already a dependency of ComfyUI
  • Pillow also already used by ComfyUI for image processing

Usage

The default base_url parameter value targets the official OpenAI API endpoint by default but by changing it, you can also use this project with any OpenAI API compatible servers like Ollama, vLLM, TGI, etc...

Multiples images are supported as long as they are fed batched to the chat completion node.

If you want to customize the chat completion, you can chain options to modify the request. Most common options are available as predefined nodes but you can inject any key/value pair using the Extra body node.

Options nodes are available for:

  • seed
  • temperature
  • max_tokens
  • top_p
  • frequency_penalty
  • presence_penalty
  • developer_role
  • extra_body (for any other key/value pair)

Installation

ComfyUI Manager

Search for OpenAI API in the Custom Nodes Manager and install it.

Entry in the ComfyUI manager

Manually

On the github interface, click the green <> Code button and then Download ZIP. Extract the root folder of the zip file into your ComfyUI/custom_nodes directory.

Example

Example

About

Call LLM and VLM in a simple way using the OpenAI API standard within ComfyUI

Topics

Resources

License

Stars

Watchers

Forks

Contributors 2

  •  
  •