| from typing import * |
|
|
| BACKEND = 'spconv' |
| DEBUG = False |
| ATTN = 'flash_attn' |
|
|
| def __from_env(): |
| import os |
| |
| global BACKEND |
| global DEBUG |
| global ATTN |
| |
| env_sparse_backend = os.environ.get('SPARSE_BACKEND') |
| env_sparse_debug = os.environ.get('SPARSE_DEBUG') |
| env_sparse_attn = os.environ.get('SPARSE_ATTN_BACKEND') |
| if env_sparse_attn is None: |
| env_sparse_attn = os.environ.get('ATTN_BACKEND') |
|
|
| if env_sparse_backend is not None and env_sparse_backend in ['spconv', 'torchsparse']: |
| BACKEND = env_sparse_backend |
| if env_sparse_debug is not None: |
| DEBUG = env_sparse_debug == '1' |
| if env_sparse_attn is not None and env_sparse_attn in ['xformers', 'flash_attn']: |
| ATTN = env_sparse_attn |
| |
| print(f"[SPARSE] Backend: {BACKEND}, Attention: {ATTN}") |
| |
|
|
| __from_env() |
| |
|
|
| def set_backend(backend: Literal['spconv', 'torchsparse']): |
| global BACKEND |
| BACKEND = backend |
|
|
| def set_debug(debug: bool): |
| global DEBUG |
| DEBUG = debug |
|
|
| def set_attn(attn: Literal['xformers', 'flash_attn']): |
| global ATTN |
| ATTN = attn |
| |
| |
| import importlib |
|
|
| __attributes = { |
| 'SparseTensor': 'basic', |
| 'sparse_batch_broadcast': 'basic', |
| 'sparse_batch_op': 'basic', |
| 'sparse_cat': 'basic', |
| 'sparse_unbind': 'basic', |
| 'SparseGroupNorm': 'norm', |
| 'SparseLayerNorm': 'norm', |
| 'SparseGroupNorm32': 'norm', |
| 'SparseLayerNorm32': 'norm', |
| 'SparseReLU': 'nonlinearity', |
| 'SparseSiLU': 'nonlinearity', |
| 'SparseGELU': 'nonlinearity', |
| 'SparseActivation': 'nonlinearity', |
| 'SparseLinear': 'linear', |
| 'sparse_scaled_dot_product_attention': 'attention', |
| 'SerializeMode': 'attention', |
| 'sparse_serialized_scaled_dot_product_self_attention': 'attention', |
| 'sparse_windowed_scaled_dot_product_self_attention': 'attention', |
| 'SparseMultiHeadAttention': 'attention', |
| 'SparseConv3d': 'conv', |
| 'SparseInverseConv3d': 'conv', |
| 'SparseDownsample': 'spatial', |
| 'SparseUpsample': 'spatial', |
| 'SparseSubdivide' : 'spatial' |
| } |
|
|
| __submodules = ['transformer'] |
|
|
| __all__ = list(__attributes.keys()) + __submodules |
|
|
| def __getattr__(name): |
| if name not in globals(): |
| if name in __attributes: |
| module_name = __attributes[name] |
| module = importlib.import_module(f".{module_name}", __name__) |
| globals()[name] = getattr(module, name) |
| elif name in __submodules: |
| module = importlib.import_module(f".{name}", __name__) |
| globals()[name] = module |
| else: |
| raise AttributeError(f"module {__name__} has no attribute {name}") |
| return globals()[name] |
|
|
|
|
| |
| if __name__ == '__main__': |
| from .basic import * |
| from .norm import * |
| from .nonlinearity import * |
| from .linear import * |
| from .attention import * |
| from .conv import * |
| from .spatial import * |
| import transformer |
|
|