Skip to content

Commit 374ee28

Browse files
authored
[Frontend] Remove custom_cache_manager (#13791)
Signed-off-by: fulvius31 <[email protected]>
1 parent a4d8366 commit 374ee28

File tree

3 files changed

+1
-71
lines changed

3 files changed

+1
-71
lines changed

vllm/executor/multiproc_worker_utils.py

-8
Original file line numberDiff line numberDiff line change
@@ -16,12 +16,8 @@
1616

1717
from vllm.config import VllmConfig
1818
from vllm.logger import init_logger
19-
from vllm.triton_utils.importing import HAS_TRITON
2019
from vllm.utils import _check_multiproc_method, get_mp_context, run_method
2120

22-
if HAS_TRITON:
23-
from vllm.triton_utils import maybe_set_triton_cache_manager
24-
2521
logger = init_logger(__name__)
2622

2723
T = TypeVar('T')
@@ -314,7 +310,3 @@ def set_multiprocessing_worker_envs(parallel_config):
314310
current_parallelism, default_omp_num_threads)
315311
os.environ["OMP_NUM_THREADS"] = str(default_omp_num_threads)
316312
torch.set_num_threads(default_omp_num_threads)
317-
318-
# workaround for https://github.com/vllm-project/vllm/issues/6103
319-
if HAS_TRITON and parallel_config.world_size > 1:
320-
maybe_set_triton_cache_manager()

vllm/triton_utils/__init__.py

+1-8
Original file line numberDiff line numberDiff line change
@@ -2,11 +2,4 @@
22

33
from vllm.triton_utils.importing import HAS_TRITON
44

5-
__all__ = ["HAS_TRITON"]
6-
7-
if HAS_TRITON:
8-
9-
from vllm.triton_utils.custom_cache_manager import (
10-
maybe_set_triton_cache_manager)
11-
12-
__all__ += ["maybe_set_triton_cache_manager"]
5+
__all__ = ["HAS_TRITON"]

vllm/triton_utils/custom_cache_manager.py

-55
This file was deleted.

0 commit comments

Comments
 (0)