ai-content-maker/.venv/Lib/site-packages/numba/cuda/testing.py

195 lines
6.0 KiB
Python
Raw Normal View History

2024-05-03 04:18:51 +03:00
import os
import platform
import shutil
from numba.tests.support import SerialMixin
from numba.cuda.cuda_paths import get_conda_ctk
from numba.cuda.cudadrv import driver, devices, libs
from numba.core import config
from numba.tests.support import TestCase
from pathlib import Path
import unittest
numba_cuda_dir = Path(__file__).parent
test_data_dir = numba_cuda_dir / 'tests' / 'data'
class CUDATestCase(SerialMixin, TestCase):
"""
For tests that use a CUDA device. Test methods in a CUDATestCase must not
be run out of module order, because the ContextResettingTestCase may reset
the context and destroy resources used by a normal CUDATestCase if any of
its tests are run between tests from a CUDATestCase.
"""
def setUp(self):
self._low_occupancy_warnings = config.CUDA_LOW_OCCUPANCY_WARNINGS
self._warn_on_implicit_copy = config.CUDA_WARN_ON_IMPLICIT_COPY
# Disable warnings about low gpu utilization in the test suite
config.CUDA_LOW_OCCUPANCY_WARNINGS = 0
# Disable warnings about host arrays in the test suite
config.CUDA_WARN_ON_IMPLICIT_COPY = 0
def tearDown(self):
config.CUDA_LOW_OCCUPANCY_WARNINGS = self._low_occupancy_warnings
config.CUDA_WARN_ON_IMPLICIT_COPY = self._warn_on_implicit_copy
class ContextResettingTestCase(CUDATestCase):
"""
For tests where the context needs to be reset after each test. Typically
these inspect or modify parts of the context that would usually be expected
to be internal implementation details (such as the state of allocations and
deallocations, etc.).
"""
def tearDown(self):
super().tearDown()
from numba.cuda.cudadrv.devices import reset
reset()
def ensure_supported_ccs_initialized():
from numba.cuda import is_available as cuda_is_available
from numba.cuda.cudadrv import nvvm
if cuda_is_available():
# Ensure that cudart.so is loaded and the list of supported compute
# capabilities in the nvvm module is populated before a fork. This is
# needed because some compilation tests don't require a CUDA context,
# but do use NVVM, and it is required that libcudart.so should be
# loaded before a fork (note that the requirement is not explicitly
# documented).
nvvm.get_supported_ccs()
def skip_on_cudasim(reason):
"""Skip this test if running on the CUDA simulator"""
return unittest.skipIf(config.ENABLE_CUDASIM, reason)
def skip_unless_cudasim(reason):
"""Skip this test if running on CUDA hardware"""
return unittest.skipUnless(config.ENABLE_CUDASIM, reason)
def skip_unless_conda_cudatoolkit(reason):
"""Skip test if the CUDA toolkit was not installed by Conda"""
return unittest.skipUnless(get_conda_ctk() is not None, reason)
def skip_if_external_memmgr(reason):
"""Skip test if an EMM Plugin is in use"""
return unittest.skipIf(config.CUDA_MEMORY_MANAGER != 'default', reason)
def skip_under_cuda_memcheck(reason):
return unittest.skipIf(os.environ.get('CUDA_MEMCHECK') is not None, reason)
def skip_without_nvdisasm(reason):
nvdisasm_path = shutil.which('nvdisasm')
return unittest.skipIf(nvdisasm_path is None, reason)
def skip_with_nvdisasm(reason):
nvdisasm_path = shutil.which('nvdisasm')
return unittest.skipIf(nvdisasm_path is not None, reason)
def skip_on_arm(reason):
cpu = platform.processor()
is_arm = cpu.startswith('arm') or cpu.startswith('aarch')
return unittest.skipIf(is_arm, reason)
def skip_if_cuda_includes_missing(fn):
# Skip when cuda.h is not available - generally this should indicate
# whether the CUDA includes are available or not
cuda_h = os.path.join(config.CUDA_INCLUDE_PATH, 'cuda.h')
cuda_h_file = (os.path.exists(cuda_h) and os.path.isfile(cuda_h))
reason = 'CUDA include dir not available on this system'
return unittest.skipUnless(cuda_h_file, reason)(fn)
def skip_if_mvc_enabled(reason):
"""Skip a test if Minor Version Compatibility is enabled"""
return unittest.skipIf(config.CUDA_ENABLE_MINOR_VERSION_COMPATIBILITY,
reason)
def skip_if_mvc_libraries_unavailable(fn):
libs_available = False
try:
import cubinlinker # noqa: F401
import ptxcompiler # noqa: F401
libs_available = True
except ImportError:
pass
return unittest.skipUnless(libs_available,
"Requires cubinlinker and ptxcompiler")(fn)
def cc_X_or_above(major, minor):
if not config.ENABLE_CUDASIM:
cc = devices.get_context().device.compute_capability
return cc >= (major, minor)
else:
return True
def skip_unless_cc_50(fn):
return unittest.skipUnless(cc_X_or_above(5, 0), "requires cc >= 5.0")(fn)
def skip_unless_cc_53(fn):
return unittest.skipUnless(cc_X_or_above(5, 3), "requires cc >= 5.3")(fn)
def skip_unless_cc_60(fn):
return unittest.skipUnless(cc_X_or_above(6, 0), "requires cc >= 6.0")(fn)
def skip_unless_cc_75(fn):
return unittest.skipUnless(cc_X_or_above(7, 5), "requires cc >= 7.5")(fn)
def xfail_unless_cudasim(fn):
if config.ENABLE_CUDASIM:
return fn
else:
return unittest.expectedFailure(fn)
def skip_with_cuda_python(reason):
return unittest.skipIf(driver.USE_NV_BINDING, reason)
def cudadevrt_missing():
if config.ENABLE_CUDASIM:
return False
try:
path = libs.get_cudalib('cudadevrt', static=True)
libs.check_static_lib(path)
except FileNotFoundError:
return True
return False
def skip_if_cudadevrt_missing(fn):
return unittest.skipIf(cudadevrt_missing(), 'cudadevrt missing')(fn)
class ForeignArray(object):
"""
Class for emulating an array coming from another library through the CUDA
Array interface. This just hides a DeviceNDArray so that it doesn't look
like a DeviceNDArray.
"""
def __init__(self, arr):
self._arr = arr
self.__cuda_array_interface__ = arr.__cuda_array_interface__