Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FR] Support OpenAI compatible API endpoint for local AI inference #7686

Open
rampa3 opened this issue Apr 4, 2025 · 0 comments
Open

[FR] Support OpenAI compatible API endpoint for local AI inference #7686

rampa3 opened this issue Apr 4, 2025 · 0 comments

Comments

@rampa3
Copy link

rampa3 commented Apr 4, 2025

Description

I suggest adding support for OpenAI compatible API endpoint for local AI inference. Main reasoning is, that while Ollama support is useful for some users, majority of the target audience of local AI users will likely already have a local AI inference API running for usage with other apps, and unless they were Ollama users since the beginning, their local API will be OAI compatible endpoint, as it is the most widespread AI API type. I think incompatibility with most common AI API type is a major obstacle slowing down adoption of the newly added free local AI option.

Impact

  • current users of local inference APIs will be able to use their current inference API without need of installing secondary API and installing models twice, as Ollama is not able to share its model library with other inference server solutions
  • new users of local inference APIs will be able to pick more user friendly server apps, serving more widely used type of inference API, and will be able to reuse the API server used with AppFlowy with other apps

Additional Context

Links to related AppFlowy Discord conversations (+ comments to the messages):

Extra context:

  • Ollama adds OpenAI compatible API to be compatible with software not exclusively written for it:
    https://ollama.com/blog/openai-compatibility
    Quote from the linked blog post: "Ollama now has built-in compatibility with the OpenAI Chat Completions API, making it possible to use more tooling and applications with Ollama locally."

Personal case to illustrate the storage constraints problem:

As a user of local AI models, I was looking forward to AppFlowy's local AI implementation since announcement, but it being limited only to Ollama as inference API server is sadly a dealbreaker, as I already run an instance of LocalAI API, and don't want to run two APIs side-by-side due to storage constraints. I have quite a big amount of various models installed in the LocalAI instance, which also means that the model folder for it is quite big. Having to duplicately install my text generation models and embedders in Ollama would take up the space for these models twice, once in LocalAI for OpenAI compatible tools, once in Ollama for AppFlowy. Doing this would mean having at minimum (counting only one text model + embedder) around 5 GiBs of duplicate files in my storage, cutting away from free space for installing apps, documents storage and system files.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant