# Copyright © 2023 Apple Inc. from mlx.nn.layers.activations import ( CELU, ELU, GELU, SELU, Hardswish, LeakyReLU, LogSigmoid, LogSoftmax, Mish, PReLU, ReLU, ReLU6, SiLU, Softmax, Softplus, Softsign, Step, Tanh, celu, elu, gelu, gelu_approx, gelu_fast_approx, hardswish, leaky_relu, log_sigmoid, log_softmax, mish, prelu, relu, relu6, selu, silu, softmax, softplus, softsign, step, tanh, ) from mlx.nn.layers.base import Module from mlx.nn.layers.containers import Sequential from mlx.nn.layers.convolution import Conv1d, Conv2d from mlx.nn.layers.dropout import Dropout, Dropout2d, Dropout3d from mlx.nn.layers.embedding import Embedding from mlx.nn.layers.linear import Bilinear, Identity, Linear from mlx.nn.layers.normalization import ( BatchNorm, GroupNorm, InstanceNorm, LayerNorm, RMSNorm, ) from mlx.nn.layers.positional_encoding import ALiBi, RoPE, SinusoidalPositionalEncoding from mlx.nn.layers.quantized import QuantizedLinear from mlx.nn.layers.transformer import ( MultiHeadAttention, Transformer, TransformerEncoder, TransformerEncoderLayer, )