ScanQA_voxel_data / preprocessing /ScanNet200 /preprocess_scannet200.py
Yezixiao's picture
Upload folder using huggingface_hub
84d92e1 verified
import warnings
warnings.filterwarnings("ignore", category=DeprecationWarning)
import sys
import os
import argparse
import glob
import json
from concurrent.futures import ProcessPoolExecutor
# Load external constants
from scannet200_constants import *
from scannet200_splits import *
from utils import *
CLOUD_FILE_PFIX = '_vh_clean_2'
SEGMENTS_FILE_PFIX = '.0.010000.segs.json'
AGGREGATIONS_FILE_PFIX = '.aggregation.json'
CLASS_IDs = VALID_CLASS_IDS_200
_OUTPUT_ROOT = ''
_TRAIN_SCENES = set()
_VAL_SCENES = set()
_VOXEL_SIZE = 0.2
_NORMALIZE = False
_LABEL_MAP = {}
def init_worker(output_root, train_scenes, val_scenes, voxel_size, normalize, label_map):
global _OUTPUT_ROOT, _TRAIN_SCENES, _VAL_SCENES, _VOXEL_SIZE, _NORMALIZE, _LABEL_MAP
_OUTPUT_ROOT = output_root
_TRAIN_SCENES = set(train_scenes)
_VAL_SCENES = set(val_scenes)
_VOXEL_SIZE = voxel_size
_NORMALIZE = normalize
_LABEL_MAP = label_map
def normalize_pointcloud(points):
centered = points - np.mean(points, axis=0, keepdims=True)
scale = np.max(np.linalg.norm(centered, axis=1))
if scale < 1e-8:
return centered
return centered / scale
def handle_process(scene_path):
scene_id = os.path.basename(scene_path)
mesh_path = os.path.join(scene_path, f'{scene_id}{CLOUD_FILE_PFIX}.ply')
segments_file = os.path.join(scene_path, f'{scene_id}{CLOUD_FILE_PFIX}{SEGMENTS_FILE_PFIX}')
aggregations_file = os.path.join(scene_path, f'{scene_id}{AGGREGATIONS_FILE_PFIX}')
info_file = os.path.join(scene_path, f'{scene_id}.txt')
if _NORMALIZE:
norm_suffix = '_normalized'
else:
norm_suffix = ''
if scene_id in _TRAIN_SCENES:
output_file = os.path.join(_OUTPUT_ROOT, 'train', f'{scene_id}{norm_suffix}.ply')
voxel_output_file = os.path.join(_OUTPUT_ROOT, 'train', f'{scene_id}_voxel_{_VOXEL_SIZE}{norm_suffix}.ply')
split_name = 'train'
elif scene_id in _VAL_SCENES:
output_file = os.path.join(_OUTPUT_ROOT, 'val', f'{scene_id}{norm_suffix}.ply')
voxel_output_file = os.path.join(_OUTPUT_ROOT, 'val', f'{scene_id}_voxel_{_VOXEL_SIZE}{norm_suffix}.ply')
split_name = 'val'
else:
output_file = os.path.join(_OUTPUT_ROOT, 'test', f'{scene_id}{norm_suffix}.ply')
voxel_output_file = os.path.join(_OUTPUT_ROOT, 'test', f'{scene_id}_voxel_{_VOXEL_SIZE}{norm_suffix}.ply')
split_name = 'test'
print('Processing: ', scene_id, 'in ', split_name)
# Rotating the mesh to axis aligned
info_dict = {}
with open(info_file) as f:
for line in f:
(key, val) = line.split(" = ")
info_dict[key] = np.fromstring(val, sep=' ')
if 'axisAlignment' not in info_dict:
rot_matrix = np.identity(4)
else:
rot_matrix = info_dict['axisAlignment'].reshape(4, 4)
mesh_data = read_plymesh(mesh_path)
if mesh_data is None:
raise ValueError(f'Empty mesh: {mesh_path}')
pointcloud, faces_array = mesh_data
# Rotate PC to axis aligned
r_points = pointcloud[:, :3].transpose()
r_points = np.append(r_points, np.ones((1, r_points.shape[1])), axis=0)
r_points = np.dot(rot_matrix, r_points)
pointcloud = np.append(r_points.transpose()[:, :3], pointcloud[:, 3:], axis=1)
if _NORMALIZE:
pointcloud[:, :3] = normalize_pointcloud(pointcloud[:, :3])
points = pointcloud[:, :3]
colors = pointcloud[:, 3:6]
# Load segments file
with open(segments_file) as f:
segments = json.load(f)
seg_indices = np.array(segments['segIndices'])
# Load Aggregations file
with open(aggregations_file) as f:
aggregation = json.load(f)
seg_groups = np.array(aggregation['segGroups'])
# Generate new labels
labelled_pc = np.zeros((pointcloud.shape[0], 1))
instance_ids = np.zeros((pointcloud.shape[0], 1))
for group in seg_groups:
p_inds, label_id = point_indices_from_group(seg_indices, group, _LABEL_MAP, CLASS_IDs)
labelled_pc[p_inds] = label_id
instance_ids[p_inds] = group['id']
labelled_pc = labelled_pc.astype(int)
instance_ids = instance_ids.astype(int)
# Concatenate with original cloud
processed_vertices = np.hstack((pointcloud[:, :6], labelled_pc, instance_ids))
if (np.any(np.isnan(processed_vertices)) or not np.all(np.isfinite(processed_vertices))):
raise ValueError('nan')
# Save processed mesh
save_plymesh(processed_vertices, faces_array, output_file, with_label=True, verbose=False)
# Uncomment the following lines if saving the output in voxelized point cloud
quantized_points, quantized_scene_colors, quantized_labels, quantized_instances = voxelize_pointcloud(
points,
colors,
labelled_pc,
instance_ids,
faces_array,
voxel_size=_VOXEL_SIZE,
)
quantized_pc = np.hstack((quantized_points, quantized_scene_colors, quantized_labels, quantized_instances))
save_plymesh(quantized_pc, faces=None, filename=voxel_output_file, with_label=True, verbose=False)
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('--dataset_root', required=True, help='Path to the ScanNet dataset containing scene folders')
parser.add_argument('--output_root', required=True, help='Output path where train/val folders will be located')
parser.add_argument('--label_map_file', required=True, help='path to scannetv2-labels.combined.tsv')
parser.add_argument('--num_workers', default=4, type=int, help='The number of parallel workers')
parser.add_argument('--train_val_splits_path', default=None, help='Where the txt files with the train/val splits live')
parser.add_argument('--voxel_size', default=0.2, type=float, help='Size of the voxel for voxelization')
parser.add_argument('--normalize_pointcloud', action='store_true', help='Normalize each scene point cloud to a unit sphere after axis alignment')
config = parser.parse_args()
# Load label map
labels_pd = pd.read_csv(config.label_map_file, sep='\t', header=0)
label_map = dict(zip(labels_pd['raw_category'], labels_pd['id']))
# Load train/val splits
with open(config.train_val_splits_path + '/scannetv2_train.txt') as train_file:
train_scenes = train_file.read().splitlines()
with open(config.train_val_splits_path + '/scannetv2_val.txt') as val_file:
val_scenes = val_file.read().splitlines()
# Create output directories
train_output_dir = os.path.join(config.output_root, 'train')
if not os.path.exists(train_output_dir):
os.makedirs(train_output_dir)
val_output_dir = os.path.join(config.output_root, 'val')
if not os.path.exists(val_output_dir):
os.makedirs(val_output_dir)
test_output_dir = os.path.join(config.output_root, 'test')
if not os.path.exists(test_output_dir):
os.makedirs(test_output_dir)
# Load scene paths
scene_paths = sorted(glob.glob(config.dataset_root + '/*'))
# Preprocess data.
print('Processing scenes...')
with ProcessPoolExecutor(
max_workers=config.num_workers,
initializer=init_worker,
initargs=(
config.output_root,
train_scenes,
val_scenes,
config.voxel_size,
config.normalize_pointcloud,
label_map,
),
) as pool:
_ = list(pool.map(handle_process, scene_paths))