| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| #include "pass_level1.h" |
|
|
| #include "../utils.h" |
|
|
| namespace pnnx { |
|
|
| class BatchNorm1d : public FuseModulePass |
| { |
| public: |
| const char* match_type_str() const |
| { |
| return "__torch__.torch.nn.modules.batchnorm.BatchNorm1d"; |
| } |
|
|
| const char* type_str() const |
| { |
| return "nn.BatchNorm1d"; |
| } |
|
|
| void write(Operator* op, const std::shared_ptr<torch::jit::Graph>& graph, const torch::jit::Module& mod) const |
| { |
| const torch::jit::Node* bn = find_node_by_kind(graph, "aten::batch_norm"); |
|
|
| const auto& running_mean = mod.attr("running_mean").toTensor(); |
| const auto& running_var = mod.attr("running_var").toTensor(); |
|
|
| op->params["num_features"] = running_mean.size(0); |
| op->params["eps"] = bn->namedInput("eps"); |
| op->params["affine"] = mod.hasattr("weight") && mod.hasattr("bias"); |
|
|
| op->attrs["running_mean"] = running_mean; |
| op->attrs["running_var"] = running_var; |
| if (mod.hasattr("weight") && mod.hasattr("bias")) |
| { |
| op->attrs["weight"] = mod.attr("weight").toTensor(); |
| op->attrs["bias"] = mod.attr("bias").toTensor(); |
| } |
| } |
| }; |
|
|
| REGISTER_GLOBAL_PNNX_FUSE_MODULE_PASS(BatchNorm1d) |
|
|
| } |
|
|