1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677 |
- """Utils."""
- from os import path
- import enum
- import socket
- from platform import uname
- import uuid
- import psutil
- import torch
- from aphrodite._C import cuda_utils
- class Device(enum.Enum):
- GPU = enum.auto()
- CPU = enum.auto()
- class Counter:
- def __init__(self, start: int = 0) -> None:
- self.counter = start
- def __next__(self) -> int:
- i = self.counter
- self.counter += 1
- return i
- def reset(self) -> None:
- self.counter = 0
- def is_hip() -> bool:
- return torch.version.hip is not None
- def get_max_shared_memory_bytes(gpu: int = 0) -> int:
- """Returns the maximum shared memory per thread block in bytes."""
- # https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html
- # pylint: disable=invalid-name
- cudaDevAttrMaxSharedMemoryPerBlockOptin = 97 if not is_hip() else 74
- max_shared_mem = cuda_utils.get_device_attribute(
- cudaDevAttrMaxSharedMemoryPerBlockOptin, gpu)
- return int(max_shared_mem)
- def get_cpu_memory() -> int:
- """Returns the total CPU memory of the node or container in bytes."""
- memory_limit = psutil.virtual_memory().total
- for limit_file in [
- "/sys/fs/cgroup/memory/memory.limit_in_bytes", # v1
- "/sys/fs/cgroup/memory.max" # v2
- ]:
- if path.exists(limit_file):
- with open(limit_file) as f:
- content = f.read().strip()
- if content.isnumeric(): # v2 can have "max" as limit
- memory_limit = min(memory_limit, int(content))
- return memory_limit
- def random_uuid() -> str:
- return str(uuid.uuid4().hex)
- def in_wsl() -> bool:
- # Reference: https://github.com/microsoft/WSL/issues/4071
- return "microsoft" in " ".join(uname()).lower()
- def get_open_port():
- with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:
- s.bind(("", 0))
- return s.getsockname()[1]
|