229 lines
5.7 KiB
Python
229 lines
5.7 KiB
Python
from .backends import (
|
|
CupyOps,
|
|
MPSOps,
|
|
NumpyOps,
|
|
Ops,
|
|
get_current_ops,
|
|
get_ops,
|
|
set_current_ops,
|
|
set_gpu_allocator,
|
|
use_ops,
|
|
use_pytorch_for_gpu_memory,
|
|
use_tensorflow_for_gpu_memory,
|
|
)
|
|
from .compat import enable_mxnet, enable_tensorflow, has_cupy
|
|
from .config import Config, ConfigValidationError, registry
|
|
from .initializers import (
|
|
configure_normal_init,
|
|
glorot_uniform_init,
|
|
normal_init,
|
|
uniform_init,
|
|
zero_init,
|
|
)
|
|
from .layers import (
|
|
LSTM,
|
|
CauchySimilarity,
|
|
ClippedLinear,
|
|
Dish,
|
|
Dropout,
|
|
Embed,
|
|
Gelu,
|
|
HardSigmoid,
|
|
HardSwish,
|
|
HardSwishMobilenet,
|
|
HardTanh,
|
|
HashEmbed,
|
|
LayerNorm,
|
|
Linear,
|
|
Logistic,
|
|
Maxout,
|
|
Mish,
|
|
MultiSoftmax,
|
|
MXNetWrapper,
|
|
ParametricAttention,
|
|
ParametricAttention_v2,
|
|
PyTorchLSTM,
|
|
PyTorchRNNWrapper,
|
|
PyTorchWrapper,
|
|
PyTorchWrapper_v2,
|
|
PyTorchWrapper_v3,
|
|
Relu,
|
|
ReluK,
|
|
Sigmoid,
|
|
Softmax,
|
|
Softmax_v2,
|
|
SparseLinear,
|
|
SparseLinear_v2,
|
|
Swish,
|
|
TensorFlowWrapper,
|
|
TorchScriptWrapper_v1,
|
|
add,
|
|
array_getitem,
|
|
bidirectional,
|
|
chain,
|
|
clone,
|
|
concatenate,
|
|
expand_window,
|
|
keras_subclass,
|
|
list2array,
|
|
list2padded,
|
|
list2ragged,
|
|
map_list,
|
|
noop,
|
|
padded2list,
|
|
premap_ids,
|
|
pytorch_to_torchscript_wrapper,
|
|
ragged2list,
|
|
reduce_first,
|
|
reduce_last,
|
|
reduce_max,
|
|
reduce_mean,
|
|
reduce_sum,
|
|
remap_ids,
|
|
remap_ids_v2,
|
|
residual,
|
|
resizable,
|
|
siamese,
|
|
sigmoid_activation,
|
|
softmax_activation,
|
|
strings2arrays,
|
|
tuplify,
|
|
uniqued,
|
|
with_array,
|
|
with_array2d,
|
|
with_cpu,
|
|
with_debug,
|
|
with_flatten,
|
|
with_flatten_v2,
|
|
with_getitem,
|
|
with_list,
|
|
with_nvtx_range,
|
|
with_padded,
|
|
with_ragged,
|
|
with_reshape,
|
|
with_signpost_interval,
|
|
)
|
|
from .loss import (
|
|
CategoricalCrossentropy,
|
|
CosineDistance,
|
|
L2Distance,
|
|
SequenceCategoricalCrossentropy,
|
|
)
|
|
from .model import (
|
|
Model,
|
|
change_attr_values,
|
|
deserialize_attr,
|
|
serialize_attr,
|
|
set_dropout_rate,
|
|
wrap_model_recursive,
|
|
)
|
|
from .optimizers import SGD, Adam, Optimizer, RAdam
|
|
from .schedules import (
|
|
compounding,
|
|
constant,
|
|
constant_then,
|
|
cyclic_triangular,
|
|
decaying,
|
|
slanted_triangular,
|
|
warmup_linear,
|
|
)
|
|
from .shims import (
|
|
MXNetShim,
|
|
PyTorchGradScaler,
|
|
PyTorchShim,
|
|
Shim,
|
|
TensorFlowShim,
|
|
TorchScriptShim,
|
|
keras_model_fns,
|
|
maybe_handshake_model,
|
|
)
|
|
from .types import ArgsKwargs, Padded, Ragged, Unserializable
|
|
from .util import (
|
|
DataValidationError,
|
|
data_validation,
|
|
fix_random_seed,
|
|
get_array_module,
|
|
get_torch_default_device,
|
|
get_width,
|
|
is_cupy_array,
|
|
mxnet2xp,
|
|
prefer_gpu,
|
|
require_cpu,
|
|
require_gpu,
|
|
set_active_gpu,
|
|
tensorflow2xp,
|
|
to_categorical,
|
|
to_numpy,
|
|
torch2xp,
|
|
xp2mxnet,
|
|
xp2tensorflow,
|
|
xp2torch,
|
|
)
|
|
|
|
# fmt: off
|
|
__all__ = [
|
|
# .config
|
|
"Config", "registry", "ConfigValidationError",
|
|
# .initializers
|
|
"normal_init", "uniform_init", "glorot_uniform_init", "zero_init",
|
|
"configure_normal_init",
|
|
# .loss
|
|
"CategoricalCrossentropy", "L2Distance", "CosineDistance",
|
|
"SequenceCategoricalCrossentropy",
|
|
# .model
|
|
"Model", "serialize_attr", "deserialize_attr",
|
|
"set_dropout_rate", "change_attr_values", "wrap_model_recursive",
|
|
# .shims
|
|
"Shim", "PyTorchGradScaler", "PyTorchShim", "TensorFlowShim", "keras_model_fns",
|
|
"MXNetShim", "TorchScriptShim", "maybe_handshake_model",
|
|
# .optimizers
|
|
"Adam", "RAdam", "SGD", "Optimizer",
|
|
# .schedules
|
|
"cyclic_triangular", "warmup_linear", "constant", "constant_then",
|
|
"decaying", "slanted_triangular", "compounding",
|
|
# .types
|
|
"Ragged", "Padded", "ArgsKwargs", "Unserializable",
|
|
# .util
|
|
"fix_random_seed", "is_cupy_array", "set_active_gpu",
|
|
"prefer_gpu", "require_gpu", "require_cpu",
|
|
"DataValidationError", "data_validation",
|
|
"to_categorical", "get_width", "get_array_module", "to_numpy",
|
|
"torch2xp", "xp2torch", "tensorflow2xp", "xp2tensorflow", "mxnet2xp", "xp2mxnet",
|
|
"get_torch_default_device",
|
|
# .compat
|
|
"enable_mxnet",
|
|
"enable_tensorflow",
|
|
"has_cupy",
|
|
# .backends
|
|
"get_ops", "set_current_ops", "get_current_ops", "use_ops",
|
|
"Ops", "CupyOps", "MPSOps", "NumpyOps", "set_gpu_allocator",
|
|
"use_pytorch_for_gpu_memory", "use_tensorflow_for_gpu_memory",
|
|
# .layers
|
|
"Dropout", "Embed", "expand_window", "HashEmbed", "LayerNorm", "Linear",
|
|
"Maxout", "Mish", "MultiSoftmax", "Relu", "softmax_activation", "Softmax", "LSTM",
|
|
"CauchySimilarity", "ParametricAttention", "Logistic",
|
|
"resizable", "sigmoid_activation", "Sigmoid", "SparseLinear",
|
|
"ClippedLinear", "ReluK", "HardTanh", "HardSigmoid",
|
|
"Dish", "HardSwish", "HardSwishMobilenet", "Swish", "Gelu",
|
|
"PyTorchWrapper", "PyTorchRNNWrapper", "PyTorchLSTM",
|
|
"TensorFlowWrapper", "keras_subclass", "MXNetWrapper",
|
|
"PyTorchWrapper_v2", "Softmax_v2", "PyTorchWrapper_v3",
|
|
"SparseLinear_v2", "TorchScriptWrapper_v1", "ParametricAttention_v2",
|
|
|
|
"add", "bidirectional", "chain", "clone", "concatenate", "noop",
|
|
"residual", "uniqued", "siamese", "list2ragged", "ragged2list",
|
|
"map_list",
|
|
"with_array", "with_array2d",
|
|
"with_padded", "with_list", "with_ragged", "with_flatten",
|
|
"with_reshape", "with_getitem", "strings2arrays", "list2array",
|
|
"list2ragged", "ragged2list", "list2padded", "padded2list",
|
|
"remap_ids", "remap_ids_v2", "premap_ids",
|
|
"array_getitem", "with_cpu", "with_debug", "with_nvtx_range",
|
|
"with_signpost_interval",
|
|
"tuplify", "with_flatten_v2",
|
|
"pytorch_to_torchscript_wrapper",
|
|
|
|
"reduce_first", "reduce_last", "reduce_max", "reduce_mean", "reduce_sum",
|
|
]
|
|
# fmt: on
|