123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110 |
- # Copyright (c) Microsoft Corporation.
- # SPDX-License-Identifier: Apache-2.0
- # DeepSpeed Team
- import os
- import distutils.spawn
- import subprocess
- from .builder import TorchCPUOpBuilder
- class AsyncIOBuilder(TorchCPUOpBuilder):
- BUILD_VAR = "DS_BUILD_AIO"
- NAME = "async_io"
- def __init__(self):
- super().__init__(name=self.NAME)
- def absolute_name(self):
- return f'deepspeed.ops.aio.{self.NAME}_op'
- def lib_sources(self):
- src_list = [
- 'csrc/aio/py_lib/deepspeed_py_io_handle.cpp', 'csrc/aio/py_lib/deepspeed_py_aio.cpp',
- 'csrc/aio/py_lib/deepspeed_py_aio_handle.cpp', 'csrc/aio/py_lib/deepspeed_aio_thread.cpp',
- 'csrc/aio/common/deepspeed_aio_utils.cpp', 'csrc/aio/common/deepspeed_aio_common.cpp',
- 'csrc/aio/common/deepspeed_aio_types.cpp', 'csrc/aio/py_lib/deepspeed_cpu_op.cpp',
- 'csrc/aio/py_lib/deepspeed_aio_op_desc.cpp', 'csrc/aio/py_lib/deepspeed_py_copy.cpp',
- 'csrc/aio/py_lib/deepspeed_pin_tensor.cpp'
- ]
- return src_list
- def sources(self):
- return self.lib_sources() + ['csrc/aio/py_lib/py_ds_aio.cpp']
- def include_paths(self):
- import torch
- if self.build_for_cpu:
- CUDA_INCLUDE = []
- elif not self.is_rocm_pytorch():
- CUDA_INCLUDE = [os.path.join(torch.utils.cpp_extension.CUDA_HOME, "include")]
- else:
- CUDA_INCLUDE = [
- os.path.join(torch.utils.cpp_extension.ROCM_HOME, "include"),
- os.path.join(torch.utils.cpp_extension.ROCM_HOME, "include", "rocrand"),
- os.path.join(torch.utils.cpp_extension.ROCM_HOME, "include", "hiprand"),
- ]
- return ['csrc/aio/py_lib', 'csrc/aio/common'] + CUDA_INCLUDE
- def cxx_args(self):
- # -O0 for improved debugging, since performance is bound by I/O
- args = super().cxx_args()
- import torch
- TORCH_MAJOR, TORCH_MINOR = map(int, torch.__version__.split('.')[0:2])
- if not (TORCH_MAJOR >= 2 and TORCH_MINOR >= 1):
- args.remove('-std=c++17')
- args.append('-std=c++14')
- args += ['-Wall', '-O0', '-shared', '-fPIC', '-Wno-reorder']
- return args
- def extra_ldflags(self):
- if self.build_for_cpu:
- return ['-fopenmp']
- import torch.utils.cpp_extension
- CUDA_HOME = torch.utils.cpp_extension.CUDA_HOME
- CUDA_LIB64 = os.path.join(CUDA_HOME, "lib64")
- ldflags = [f'-L{CUDA_HOME}', f'-L{CUDA_LIB64}', '-laio', '-lcuda', '-lcudart']
- return ldflags
- def check_for_libaio_pkg(self):
- libs = dict(
- dpkg=["-l", "libaio-dev", "apt"],
- pacman=["-Q", "libaio", "pacman"],
- rpm=["-q", "libaio-devel", "yum"],
- )
- found = False
- for pkgmgr, data in libs.items():
- flag, lib, tool = data
- path = distutils.spawn.find_executable(pkgmgr)
- if path is not None:
- cmd = [pkgmgr, flag, lib]
- result = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
- if result.wait() == 0:
- found = True
- else:
- self.warning(f"{self.NAME}: please install the {lib} package with {tool}")
- break
- return found
- def is_compatible(self, verbose=False):
- # Check for the existence of libaio by using distutils
- # to compile and link a test program that calls io_submit,
- # which is a function provided by libaio that is used in the async_io op.
- # If needed, one can define -I and -L entries in CFLAGS and LDFLAGS
- # respectively to specify the directories for libaio.h and libaio.so.
- aio_compatible = self.has_function('io_submit', ('aio', ))
- if verbose and not aio_compatible:
- self.warning(f"{self.NAME} requires the dev libaio .so object and headers but these were not found.")
- # Check for the libaio package via known package managers
- # to print suggestions on which package to install.
- self.check_for_libaio_pkg()
- self.warning(
- "If libaio is already installed (perhaps from source), try setting the CFLAGS and LDFLAGS environment variables to where it can be found."
- )
- return super().is_compatible(verbose) and aio_compatible
|