-
Notifications
You must be signed in to change notification settings - Fork 431
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Make Warp work with Local Language Models (like Ollama models) #4339
Comments
Some local LLM tools even expose an OpenAI API-compatible server, which I believe could allow for rudimentary support with just a setting to change the OpenAI endpoint? |
Thanks for this feature request @hmdz105! To anyone else interested in this feature, please add a 👍 to the original post at the top to signal that you want this feature, and subscribe if you'd like to be notified. |
This is the feature that would bring me to use Warp |
As a warp user and system's administrator, this feature alone would make getting buy in from our CTO 100x easier. I love using Warp in my home lab / personal environments but no way in its current iteration could I even get it in my professional environment to demo a proof of value. Having the option to show case the AI integrations with local models would allow us to begin the discussion on an enterprise plan with zero data retention. |
Yes, please. Ollama is easiest. LMStudio great too. Groq (not the X model) would be rather interesting. Let people use the LLM they want to use. Warp team, this INCREASES the value of this tool greatly! Your value is NOT in being a niche gateway to popular LLMs. |
I think this is one of the important features missing in warp. 2 reasons - Most of the big companies dont let access gpt or any other LLM which captures data. Also, if they allow, there is only so many request that can be made for free. |
Would be game changer. |
vscode with LmStudio is amazing, WARP could integrate very easily! |
Consider using the keyboard shortcut action to generate prompt based on last entered command with printed output and ask it each time in order to avoid using too much CPU/data processing usage. |
This is a perfect feature for those developer who have security concerns. It's inevitable, soon enough all devices will use LLM natively just like its part of a system by default (like a RAM, CPU). Highly suggest they would add this to the roadmap, it's easy to deploy an LLM server locally nowadays. If they won't allow other services to connect to warp, maybe they can add this to the build natively: https://cortex.so/ So everyone is happy and not leaving warp desktop app just to run an llm server, its within the warp app itself. |
This shoud have been like 2nd on your list of features to implement right after plugging OpenAI API... |
I concur, I'm already running Ollama for handling other things, and ideally I'm already using qwen2.5-coder so just adding this to my terminal would mean I wouldn't need to pop out of my terminal to check something and I don't intend to use OpenAI if I can help it. |
Agreed that this option, especially at the free level, would allow for greater adoption in corporations if anything to validate the concept especially in organizations where they run their own LLM. |
This feature will make me enable AI in the app because it allows to use local AI instead of US-hosted AI(which is a huge privacy issue for me, i trust China more than US and running local is the best for me). |
Warp without this feature is a dealbreaker for me. I work with too much sensitive information to not have an option to utilize a local LLM. |
Sorry, but I will never be able to trust any cloud-based AI with any kind of sensitive information (personal or professional). Having the ability to plug the terminal into my locally ran Ollama would help me alleviate my concerns. But so far, it's unfortunately a no-go for me. |
+1 |
This seems extremely important to users, would be nice to hear from the warp team whether or not they intend to implement? If not, we can save ourselves the time watching this thread. |
ABSOLUTELY! I only run local LLMs, for security and fiscal reasons. |
Just learned about warp and have been giving it a test drive. If I could link it back to my Ollama instance, it would be a daily driver for sure! |
Running warp on a local LLM would be amazing. |
Agree with all of the above. Amazing product otherwise, but being able to integrate with a private, local LLM would make it perfect for competent privacy minded users. Leaving it as an 'Enterprise Only' feature is just a cash grab. Not cool. |
I am a novice programmer at best and I'm trying to learn by doing. I've used the free version of Warp and ran out of tokens in less than a week. I really like Warp and it has helped me learn but since I already am paying for Github Copilot and a couple more services, I am not going to add another subscription in which the AI can only be used for one purpose. So I'm uninstalling it and just using VS Code terminal or other method where I can use my existing subscriptions or use my local llm. |
It can be used to avoid using API data limits and without processing data to server. |
Warp is great, but I will not use it until we can use local LLMs |
agreed, when working with many proprietary dataset, it's not possible to take advantage of warp at all regardless of the privacy/security policies associated with warp. |
Adding my voice to the din: this looks like a really cool product but this feature is a hard requirement. Years ago, I was initially turned away by the always-online account-required policy, but it looks like that has been updated. That said, it's a nonstarter for me without being able to target any OpenAI-compatible backend. |
Love it. Have so much shit to pay for though. |
it would be a really complete tool with that feature. |
local models please 🙏 |
with this feature I would also use this in my company |
Please add a local llm feature!! Love the program but can't afford to pay for more stuff. |
local llm will help a lot.. we are just to much paying AI stuff out there. please add the features |
It's been over a year, still they haven't added local LLM options. It's spring 2025, wake up, almost every place has local LLM options these days. This is a basic and much needed feature! |
This is a much needed feature. As mentioned above, ollama has an endpoint which should be accesible with little to no effort. Allowing to setup the AI Endpoint would be a very good approach |
Please make it a priority to add this. We've been waiting for a long time at this point. |
too long did't wait |
Just do it) |
Warp's current AI functionality relies on cloud-based language models. While powerful, this limits usage for developers with strict privacy, offline, or performance requirements. Supporting local LLMs (e.g., Ollama) would bring powerful, private, and offline AI functionality directly into the Warp experience. Benefits of Local LLM Integration for Warp:1. Enhanced Privacy & Security
2. Offline Usage
3. Speed & Latency
4. Customization & Fine-tuning
5. Ollama Ecosystem Compatibility
💬 Why This Matters:Warp has redefined what a terminal can be by adding developer-friendly features, speed, and intelligent suggestions. Adding support for local AI would make Warp truly future-proof and developer-centric—honoring the same principles of flexibility and control that made the CLI powerful in the first place. ^^Generated with the localhosted LLM Warp does not enable me to implement.^^ Now on to my personal (non-ai generated) opinion: The only way adding local LLM support doesn't make sense is if they just want to milk the AI feature in their paid plans and enabling local LLM support on the free-tier will circumvent said milking. In which case, it's just lack of creativity, because it can just as easily be part of a paid plan feature! I'd happily pay for a reasonably priced WARP plan, especially when it let's me use local LLM's. |
I'd love for support for local LLMs. |
Would love this as well! |
Discord username (optional)
No response
Describe the solution you'd like?
Due to safety concerns, many users are getting paranoid about Warp's forced login and online AI assistance. Since terminals are used to access critical data on local machines and servers, adding the ability to use local language models like Llama 2 using Ollama instead of online AI engines would be a great plus to Warp terminal emulator.
Is your feature request related to a problem? Please describe.
Mostly safety concerns when using Warp for accessing password protected systems and documents.
Additional context
No response
How important is this feature to you?
4
Warp Internal (ignore) - linear-label:39cc6478-1249-4ee7-950b-c428edfeecd1
None
The text was updated successfully, but these errors were encountered: