Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Make Warp work with Local Language Models (like Ollama models) #4339

Open
hmdz105 opened this issue Feb 26, 2024 · 41 comments
Open

Make Warp work with Local Language Models (like Ollama models) #4339

hmdz105 opened this issue Feb 26, 2024 · 41 comments
Labels
Feature requests Feature Requests

Comments

@hmdz105
Copy link

hmdz105 commented Feb 26, 2024

Discord username (optional)

No response

Describe the solution you'd like?

Due to safety concerns, many users are getting paranoid about Warp's forced login and online AI assistance. Since terminals are used to access critical data on local machines and servers, adding the ability to use local language models like Llama 2 using Ollama instead of online AI engines would be a great plus to Warp terminal emulator.

Is your feature request related to a problem? Please describe.

Mostly safety concerns when using Warp for accessing password protected systems and documents.

Additional context

No response

How important is this feature to you?

4

Warp Internal (ignore) - linear-label:39cc6478-1249-4ee7-950b-c428edfeecd1

None

@Absolucy
Copy link

Absolucy commented Mar 9, 2024

Some local LLM tools even expose an OpenAI API-compatible server, which I believe could allow for rudimentary support with just a setting to change the OpenAI endpoint?

@dannyneira
Copy link
Member

dannyneira commented Apr 3, 2024

Thanks for this feature request @hmdz105!

To anyone else interested in this feature, please add a 👍 to the original post at the top to signal that you want this feature, and subscribe if you'd like to be notified.

@dannyneira dannyneira marked this as a duplicate of #3779 Apr 3, 2024
@dannyneira dannyneira closed this as not planned Won't fix, can't repro, duplicate, stale Apr 3, 2024
@dannyneira dannyneira marked this as not a duplicate of #3779 Jan 17, 2025
@dannyneira dannyneira reopened this Jan 17, 2025
@dannyneira dannyneira pinned this issue Jan 17, 2025
@kn0wmad
Copy link

kn0wmad commented Jan 19, 2025

This is the feature that would bring me to use Warp

@Ryj3l
Copy link

Ryj3l commented Jan 20, 2025

As a warp user and system's administrator, this feature alone would make getting buy in from our CTO 100x easier. I love using Warp in my home lab / personal environments but no way in its current iteration could I even get it in my professional environment to demo a proof of value. Having the option to show case the AI integrations with local models would allow us to begin the discussion on an enterprise plan with zero data retention.

@kentlangley
Copy link

Yes, please. Ollama is easiest. LMStudio great too. Groq (not the X model) would be rather interesting. Let people use the LLM they want to use. Warp team, this INCREASES the value of this tool greatly! Your value is NOT in being a niche gateway to popular LLMs.

@dannyneira dannyneira unpinned this issue Jan 24, 2025
@adityagaddhyanphonepe
Copy link

I think this is one of the important features missing in warp. 2 reasons - Most of the big companies dont let access gpt or any other LLM which captures data. Also, if they allow, there is only so many request that can be made for free.

@beamop
Copy link

beamop commented Jan 30, 2025

Would be game changer.

@lucieleblanc lucieleblanc marked this as a duplicate of #5735 Jan 30, 2025
@lucieleblanc lucieleblanc mentioned this issue Jan 30, 2025
1 task
@mbalasko
Copy link

mbalasko commented Feb 5, 2025

vscode with LmStudio is amazing, WARP could integrate very easily!

@lulunac27a
Copy link

Consider using the keyboard shortcut action to generate prompt based on last entered command with printed output and ask it each time in order to avoid using too much CPU/data processing usage.

@rodrigoisabelo
Copy link

This is a perfect feature for those developer who have security concerns. It's inevitable, soon enough all devices will use LLM natively just like its part of a system by default (like a RAM, CPU).

Highly suggest they would add this to the roadmap, it's easy to deploy an LLM server locally nowadays.

If they won't allow other services to connect to warp, maybe they can add this to the build natively: https://cortex.so/

So everyone is happy and not leaving warp desktop app just to run an llm server, its within the warp app itself.

@Semt3x
Copy link

Semt3x commented Feb 25, 2025

This shoud have been like 2nd on your list of features to implement right after plugging OpenAI API...
There is no way anybody would run such a tool as an agent over a production server over ssh
You can put as much safechecks as you want, you can't prove that OpenAi or other company won't save anything that goes through the endpoint, I'm also pretty sure that everything is routed through your own enpoints which are prone to attacks as well. If I'm not trusting my paid VPN nor my company one, there is no way I'd could trust an outsourced/offshore LLM messing with critical data. Regardless of what it can actually run on its own or not.

@atomicangel
Copy link

I concur, I'm already running Ollama for handling other things, and ideally I'm already using qwen2.5-coder so just adding this to my terminal would mean I wouldn't need to pop out of my terminal to check something and I don't intend to use OpenAI if I can help it.

@JuliusPIV
Copy link

Agreed that this option, especially at the free level, would allow for greater adoption in corporations if anything to validate the concept especially in organizations where they run their own LLM.

@PtDragon
Copy link

This feature will make me enable AI in the app because it allows to use local AI instead of US-hosted AI(which is a huge privacy issue for me, i trust China more than US and running local is the best for me).
Yes, i seen no training policy but the data can be used not only for training but for example for US government organizations to extract code parts and find vulnerabilities that can be secretly exploited by them.

@aidanjsamuels
Copy link

Warp without this feature is a dealbreaker for me. I work with too much sensitive information to not have an option to utilize a local LLM.

@jeefuji
Copy link

jeefuji commented Feb 27, 2025

Sorry, but I will never be able to trust any cloud-based AI with any kind of sensitive information (personal or professional).
I've checked the network logs, and I see calls to your GraphQL API with pretty much the output of all my previous commands in it...

Having the ability to plug the terminal into my locally ran Ollama would help me alleviate my concerns. But so far, it's unfortunately a no-go for me.

@e-p-s
Copy link

e-p-s commented Feb 27, 2025

+1

@kn0wmad
Copy link

kn0wmad commented Feb 27, 2025

This seems extremely important to users, would be nice to hear from the warp team whether or not they intend to implement? If not, we can save ourselves the time watching this thread.

@MrMxyzptlk
Copy link

ABSOLUTELY! I only run local LLMs, for security and fiscal reasons.

@Zelixo
Copy link

Zelixo commented Feb 27, 2025

Just learned about warp and have been giving it a test drive. If I could link it back to my Ollama instance, it would be a daily driver for sure!

@codinkai
Copy link

Running warp on a local LLM would be amazing.

@vZeroG
Copy link

vZeroG commented Feb 28, 2025

Agree with all of the above. Amazing product otherwise, but being able to integrate with a private, local LLM would make it perfect for competent privacy minded users.

Leaving it as an 'Enterprise Only' feature is just a cash grab. Not cool.

@lukesdaddy
Copy link

I am a novice programmer at best and I'm trying to learn by doing. I've used the free version of Warp and ran out of tokens in less than a week. I really like Warp and it has helped me learn but since I already am paying for Github Copilot and a couple more services, I am not going to add another subscription in which the AI can only be used for one purpose. So I'm uninstalling it and just using VS Code terminal or other method where I can use my existing subscriptions or use my local llm.

@lulunac27a
Copy link

It can be used to avoid using API data limits and without processing data to server.

@TheLazyLemur
Copy link

Warp is great, but I will not use it until we can use local LLMs

@faux123
Copy link

faux123 commented Mar 6, 2025

agreed, when working with many proprietary dataset, it's not possible to take advantage of warp at all regardless of the privacy/security policies associated with warp.

@jbohanon
Copy link

jbohanon commented Mar 7, 2025

Adding my voice to the din: this looks like a really cool product but this feature is a hard requirement. Years ago, I was initially turned away by the always-online account-required policy, but it looks like that has been updated. That said, it's a nonstarter for me without being able to target any OpenAI-compatible backend.

@ProJamKiller
Copy link

Love it. Have so much shit to pay for though.

@pflajszer
Copy link

it would be a really complete tool with that feature.

@gagichce
Copy link

local models please 🙏

@ultimatebash
Copy link

with this feature I would also use this in my company

@BrianTheMessiah
Copy link

Please add a local llm feature!! Love the program but can't afford to pay for more stuff.

@ribonred
Copy link

local llm will help a lot.. we are just to much paying AI stuff out there. please add the features

@aravhawk
Copy link

aravhawk commented Mar 21, 2025

It's been over a year, still they haven't added local LLM options. It's spring 2025, wake up, almost every place has local LLM options these days. This is a basic and much needed feature!

@Roukurai
Copy link

This is a much needed feature. As mentioned above, ollama has an endpoint which should be accesible with little to no effort. Allowing to setup the AI Endpoint would be a very good approach

@dannyneira dannyneira pinned this issue Mar 26, 2025
@usr-was-taken
Copy link

Please make it a priority to add this. We've been waiting for a long time at this point.

@anhvth
Copy link

anhvth commented Apr 2, 2025

too long did't wait

@aredoff
Copy link

aredoff commented Apr 2, 2025

Just do it)

@morales2k
Copy link

Warp's current AI functionality relies on cloud-based language models. While powerful, this limits usage for developers with strict privacy, offline, or performance requirements. Supporting local LLMs (e.g., Ollama) would bring powerful, private, and offline AI functionality directly into the Warp experience.

Benefits of Local LLM Integration for Warp:

1. Enhanced Privacy & Security

Many developers (especially in enterprise or regulated industries) cannot send code snippets or command history to external APIs.

Local models keep all interactions secure and fully offline.

2. Offline Usage

Warp's AI suggestions are currently unavailable without an internet connection.

With local LLM support, users can get AI-powered suggestions even when working remotely, on planes, or in secure air-gapped environments.

3. Speed & Latency

Local inference is often faster than waiting for a round trip to the cloud—especially for small queries like CLI command completions or flag explanations.

Reduces perceived latency, enhancing productivity in fast workflows.

4. Customization & Fine-tuning

Developers can fine-tune local models on internal command sets, scripts, or usage patterns.

This unlocks powerful, domain-specific completions and suggestions tailored to the team or project.

5. Ollama Ecosystem Compatibility

Ollama makes running local models easy with a single command (ollama run llama2, ollama run codellama, etc.).

Supporting Ollama would make local LLM integration seamless and approachable, without Warp needing to handle GPU/CPU details.

💬 Why This Matters:

Warp has redefined what a terminal can be by adding developer-friendly features, speed, and intelligent suggestions. Adding support for local AI would make Warp truly future-proof and developer-centric—honoring the same principles of flexibility and control that made the CLI powerful in the first place.

^^Generated with the localhosted LLM Warp does not enable me to implement.^^

Now on to my personal (non-ai generated) opinion:

The only way adding local LLM support doesn't make sense is if they just want to milk the AI feature in their paid plans and enabling local LLM support on the free-tier will circumvent said milking. In which case, it's just lack of creativity, because it can just as easily be part of a paid plan feature! I'd happily pay for a reasonably priced WARP plan, especially when it let's me use local LLM's.

@smultar
Copy link

smultar commented Apr 5, 2025

I'd love for support for local LLMs.

@steeeeev0
Copy link

Would love this as well!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Feature requests Feature Requests
Projects
None yet
Development

No branches or pull requests