vllm/vllm/utils.py

72 lines
1.6 KiB
Python
Raw Normal View History

2023-02-09 11:26:50 +00:00
import enum
import os
import socket
2023-05-23 21:39:50 -07:00
import uuid
from platform import uname
from typing import List
2023-03-22 04:45:42 +08:00
2023-05-09 15:30:12 -07:00
import psutil
2023-03-22 04:45:42 +08:00
import torch
from vllm._C import cuda_utils
2023-02-09 11:26:50 +00:00
class Device(enum.Enum):
GPU = enum.auto()
CPU = enum.auto()
class Counter:
def __init__(self, start: int = 0) -> None:
self.counter = start
2023-02-14 01:19:27 +00:00
def __next__(self) -> int:
i = self.counter
2023-02-09 11:26:50 +00:00
self.counter += 1
return i
2023-02-09 11:26:50 +00:00
def reset(self) -> None:
self.counter = 0
2023-03-22 04:45:42 +08:00
2023-03-29 14:48:56 +08:00
def is_hip() -> bool:
return torch.version.hip is not None
def get_max_shared_memory_bytes(gpu: int = 0) -> int:
"""Returns the maximum shared memory per thread block in bytes."""
# https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html
cudaDevAttrMaxSharedMemoryPerBlockOptin = 97 if not is_hip() else 74
max_shared_mem = cuda_utils.get_device_attribute(
cudaDevAttrMaxSharedMemoryPerBlockOptin, gpu)
return int(max_shared_mem)
2023-03-29 14:48:56 +08:00
def get_cpu_memory() -> int:
"""Returns the total CPU memory of the node in bytes."""
2023-03-29 14:48:56 +08:00
return psutil.virtual_memory().total
2023-05-23 21:39:50 -07:00
def random_uuid() -> str:
return str(uuid.uuid4().hex)
def in_wsl() -> bool:
# Reference: https://github.com/microsoft/WSL/issues/4071
return "microsoft" in " ".join(uname()).lower()
def get_ip() -> str:
return socket.gethostbyname(socket.gethostname())
def get_open_port() -> int:
with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:
s.bind(("", 0))
return s.getsockname()[1]
def set_cuda_visible_devices(device_ids: List[int]) -> None:
os.environ["CUDA_VISIBLE_DEVICES"] = ",".join(map(str, device_ids))