-
Notifications
You must be signed in to change notification settings - Fork 64
Shared weights whenever multiple instances #18
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
any progress? |
+1 for this. Did some benchmarking on this today. This is with 1 instance of 3 ONNX models
This is with 2 instances of 3 ONNX models
|
CC @pranavsharma does ORT provides API for doing so? Or can a ORT session be run for different inferences in parallel? |
Not fully following. What API are you looking for? I believe Triton already creates a separate session for each instance and these instances (sessions) can be used to run inferences in parallel. The drawback is that each session has its own copy of the weights thereby increasing (replicating) the memory consumption. Someone has submitted code changes to share a session between different instances. We're reviewing the changes. This should fix the memory consumption problem. |
Yes, this is what I was looking for. Sorry for not being clear in my previous question, just mumbling different ways to use a copy of weights in multiple instances that I have seen in different framework. i.e. TRT store weights in an "engine" and it can creates multiple "context" maps to the same "engine" |
@pranavsharma any progress about "Sharing a session between different instances of ONNXRuntime" ? |
I should be able to get to it this week. |
GOOD! I look forward to hearing from you soon. |
@pranavsharma any progress? |
Is there any news about sharing gpu memory? I the PR you mentioned #141, @pranavsharma ? We have to switch models regularly and sharing memory would be very beneficial. |
Also CPU
The text was updated successfully, but these errors were encountered: