Spaces:
Runtime error
Runtime error
| """ | |
| Based on: https://github.com/yanx27/Pointnet_Pointnet2_pytorch/blob/eb64fe0b4c24055559cea26299cb485dcb43d8dd/models/pointnet2_cls_ssg.py | |
| MIT License | |
| Copyright (c) 2019 benny | |
| Permission is hereby granted, free of charge, to any person obtaining a copy | |
| of this software and associated documentation files (the "Software"), to deal | |
| in the Software without restriction, including without limitation the rights | |
| to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
| copies of the Software, and to permit persons to whom the Software is | |
| furnished to do so, subject to the following conditions: | |
| The above copyright notice and this permission notice shall be included in all | |
| copies or substantial portions of the Software. | |
| THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
| IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
| FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | |
| AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
| LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
| OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | |
| SOFTWARE. | |
| """ | |
| import torch.nn as nn | |
| import torch.nn.functional as F | |
| from pointnet2_utils import PointNetSetAbstraction | |
| class get_model(nn.Module): | |
| def __init__(self, num_class, normal_channel=True, width_mult=1): | |
| super(get_model, self).__init__() | |
| self.width_mult = width_mult | |
| in_channel = 6 if normal_channel else 3 | |
| self.normal_channel = normal_channel | |
| self.sa1 = PointNetSetAbstraction( | |
| npoint=512, | |
| radius=0.2, | |
| nsample=32, | |
| in_channel=in_channel, | |
| mlp=[64 * width_mult, 64 * width_mult, 128 * width_mult], | |
| group_all=False, | |
| ) | |
| self.sa2 = PointNetSetAbstraction( | |
| npoint=128, | |
| radius=0.4, | |
| nsample=64, | |
| in_channel=128 * width_mult + 3, | |
| mlp=[128 * width_mult, 128 * width_mult, 256 * width_mult], | |
| group_all=False, | |
| ) | |
| self.sa3 = PointNetSetAbstraction( | |
| npoint=None, | |
| radius=None, | |
| nsample=None, | |
| in_channel=256 * width_mult + 3, | |
| mlp=[256 * width_mult, 512 * width_mult, 1024 * width_mult], | |
| group_all=True, | |
| ) | |
| self.fc1 = nn.Linear(1024 * width_mult, 512 * width_mult) | |
| self.bn1 = nn.BatchNorm1d(512 * width_mult) | |
| self.drop1 = nn.Dropout(0.4) | |
| self.fc2 = nn.Linear(512 * width_mult, 256 * width_mult) | |
| self.bn2 = nn.BatchNorm1d(256 * width_mult) | |
| self.drop2 = nn.Dropout(0.4) | |
| self.fc3 = nn.Linear(256 * width_mult, num_class) | |
| def forward(self, xyz, features=False): | |
| B, _, _ = xyz.shape | |
| if self.normal_channel: | |
| norm = xyz[:, 3:, :] | |
| xyz = xyz[:, :3, :] | |
| else: | |
| norm = None | |
| l1_xyz, l1_points = self.sa1(xyz, norm) | |
| l2_xyz, l2_points = self.sa2(l1_xyz, l1_points) | |
| l3_xyz, l3_points = self.sa3(l2_xyz, l2_points) | |
| x = l3_points.view(B, 1024 * self.width_mult) | |
| x = self.drop1(F.relu(self.bn1(self.fc1(x)))) | |
| result_features = self.bn2(self.fc2(x)) | |
| x = self.drop2(F.relu(result_features)) | |
| x = self.fc3(x) | |
| x = F.log_softmax(x, -1) | |
| if features: | |
| return x, l3_points, result_features | |
| else: | |
| return x, l3_points | |
| class get_loss(nn.Module): | |
| def __init__(self): | |
| super(get_loss, self).__init__() | |
| def forward(self, pred, target, trans_feat): | |
| total_loss = F.nll_loss(pred, target) | |
| return total_loss | |