repo
stringlengths
1
99
file
stringlengths
13
239
code
stringlengths
0
59.2M
file_length
int64
0
59.2M
avg_line_length
float64
0
3.34M
max_line_length
int64
0
26.7M
extension_type
stringclasses
1 value
synfeal
synfeal-main/models/depthnet.py
import torch import torch.nn as nn import torch.nn.parallel import torch.utils.data import numpy as np import torch.nn.functional as F class CNNDepth(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepth, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=5, stride=2, padding=2) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=2, padding=2) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=2) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=2) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=5, stride=2, padding=2) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.conv1(x)) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.conv2(x)) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.conv3(x)) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.conv4(x)) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.conv5(x)) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.fc1(x)) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.fc2(x)) if verbose: print('fc2 shape ' + str(x.shape)) x = F.relu(self.fc3(x)) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthLow(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthLow, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=3, stride=2) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=3, stride=2) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=3, stride=2) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=3, stride=2) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=3, stride=2) self.fc1 = nn.Linear(18432, 4096) self.fc2 = nn.Linear(4096, 1024) #self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(1024, 3) self.fc_out_rotation = nn.Linear(1024, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.conv1(x)) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.conv2(x)) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.conv3(x)) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.conv4(x)) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.conv5(x)) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.fc1(x)) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.fc2(x)) if verbose: print('fc2 shape ' + str(x.shape)) # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthDropout(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthDropout, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=5, stride=2, padding=2) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=2, padding=2) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=2) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=2) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=5, stride=2, padding=2) self.dropout1 = nn.Dropout(p=0.5) self.dropout2 = nn.Dropout(p=0.3) self.dropout3 = nn.Dropout(p=0.2) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.droupout3(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.dropout3(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.dropout3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.dropout3(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.dropout3(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.dropout1(self.fc1(x))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.dropout2(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) x = F.relu(self.dropout3(self.fc3(x))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatch(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatch, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=5, stride=2, padding=2) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=2, padding=2) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=2) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=2) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=5, stride=2, padding=2) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) self.bn4 = nn.BatchNorm2d(512) self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) self.dropout = nn.Dropout(p=0.4) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.dropout(self.bn6(self.fc1(x)))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.bn7(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) x = F.relu(self.bn8(self.fc3(x))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchK3(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchK3, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=3, stride=2, padding=1) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=3, stride=2, padding=1) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=3, stride=2, padding=1) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=3, stride=2, padding=1) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=3, stride=2, padding=1) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) self.bn4 = nn.BatchNorm2d(512) self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) self.dropout = nn.Dropout(p=0.4) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=True): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.dropout(self.bn6(self.fc1(x)))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.bn7(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) x = F.relu(self.bn8(self.fc3(x))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchLeaky(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchLeaky, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=5, stride=2, padding=2) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=2, padding=2) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=2) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=2) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=5, stride=2, padding=2) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) self.bn4 = nn.BatchNorm2d(512) self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) self.lrelu = nn.LeakyReLU(0.1) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = self.lrelu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = self.lrelu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = self.lrelu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = self.lrelu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = self.lrelu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = self.lrelu(self.dropout(self.bn6(self.fc1(x)))) if verbose: print('fc1 shape ' + str(x.shape)) x = self.lrelu(self.bn7(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) x = self.lrelu(self.bn8(self.fc3(x))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchLow(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchLow, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=32, kernel_size=3, stride=2, padding=1) self.conv2 = nn.Conv2d(in_channels=32, out_channels=64, kernel_size=3, stride=2, padding=1) self.conv3 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=3, stride=2, padding=1) self.conv4 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=3, stride=2, padding=1) self.conv5 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=3, stride=2, padding=1) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(32) self.bn2 = nn.BatchNorm2d(64) self.bn3 = nn.BatchNorm2d(128) self.bn4 = nn.BatchNorm2d(256) self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) #self.bn8 = nn.BatchNorm1d(512) #self.lrelu = nn.LeakyReLU(0.2) self.dropout = nn.Dropout(p=0.5) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) #self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(1024, 3) self.fc_out_rotation = nn.Linear(1024, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=True): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = self.lrelu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = self.lrelu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = self.lrelu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.dropout(self.bn6(self.fc1(x)))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.bn7(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) # x = self.lrelu(self.bn8(self.fc3(x))) # if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchLowL2RegLeaky(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchLowL2RegLeaky, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=3, stride=3, padding=1) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=3, stride=3, padding=1) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=3, stride=3, padding=1) #self.conv4 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=3, stride=3, padding=1) #self.conv5 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=3, stride=3, padding=1) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) #self.bn4 = nn.BatchNorm2d(256) #self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) self.lrelu = nn.LeakyReLU(0.2) #self.dropout = nn.Dropout(p=0.5) self.fc1 = nn.Linear(20736, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = self.lrelu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = self.lrelu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = self.lrelu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) # x = self.lrelu(self.bn4(self.conv4(x))) # if verbose: print('layer4 shape ' + str(x.shape)) # x = self.lrelu(self.bn5(self.conv5(x))) # if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = self.lrelu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = self.lrelu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = self.lrelu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = self.lrelu(self.bn6(self.fc1(x))) if verbose: print('fc1 shape ' + str(x.shape)) x = self.lrelu(self.bn7(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) x = self.lrelu(self.bn8(self.fc3(x))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchLowL2Reg2(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchLowL2Reg2, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=5, stride=2, padding=1) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=2, padding=1) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=1) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=1) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=5, stride=2, padding=1) #self.conv6 = nn.Conv2d(in_channels=512, out_channels=1024, kernel_size=5, stride=2, padding=1) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) self.bn4 = nn.BatchNorm2d(512) self.bn5 = nn.BatchNorm2d(512) #self.bn6 = nn.BatchNorm2d(1024) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) #self.lrelu = nn.LeakyReLU(0.2) #self.dropout = nn.Dropout(p=0.5) self.fc1 = nn.Linear(18432, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) # x = self.lrelu(self.bn6(self.conv6(x))) # if verbose: print('layer6 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = self.lrelu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = self.lrelu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = self.lrelu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.bn6(self.fc1(x))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.bn7(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) x = F.relu(self.bn8(self.fc3(x))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchDropout8(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchDropout8, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=5, stride=2, padding=2) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=2, padding=2) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=2) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=2) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=5, stride=2, padding=2) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) self.bn4 = nn.BatchNorm2d(512) self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) self.dropout = nn.Dropout(p=0.8) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.dropout(self.bn6(self.fc1(x)))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.bn7(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) x = F.relu(self.bn8(self.fc3(x))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchDropoutVar(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchDropoutVar, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=5, stride=2, padding=2) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=2, padding=2) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=2) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=2) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=5, stride=2, padding=2) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) self.bn4 = nn.BatchNorm2d(512) self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) self.dropout1 = nn.Dropout(p=0.5) self.dropout2 = nn.Dropout(p=0.3) self.dropout3 = nn.Dropout(p=0.2) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.dropout1(self.bn6(self.fc1(x)))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.dropout2(self.bn7(self.fc2(x)))) if verbose: print('fc2 shape ' + str(x.shape)) x = F.relu(self.dropout3(self.bn8(self.fc3(x)))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchDropout8Cont(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchDropout8Cont, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=5, stride=2, padding=2) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=2, padding=2) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=2) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=2) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=5, stride=2, padding=2) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) self.bn4 = nn.BatchNorm2d(512) self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) self.dropout1 = nn.Dropout(p=0.8) self.dropout2 = nn.Dropout(p=0.5) self.dropout3 = nn.Dropout(p=0.3) self.fc1 = nn.Linear(25088, 4096) self.fc2 = nn.Linear(4096, 1024) #self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(1024, 3) self.fc_out_rotation = nn.Linear(1024, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.dropout1(self.bn6(self.fc1(x)))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.dropout2(self.bn7(self.fc2(x)))) if verbose: print('fc2 shape ' + str(x.shape)) #x = F.relu(self.dropout3(self.bn8(self.fc3(x)))) #if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose class CNNDepthBatchDropout8Kernel7(nn.Module): #https://towardsdatascience.com/covolutional-neural-network-cb0883dd6529 def __init__(self): super(CNNDepthBatchDropout8Kernel7, self).__init__() # call the init constructor of the nn.Module. This way, we are only adding attributes. self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size=7, stride=2, padding=1) self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=7, stride=2, padding=1) self.conv3 = nn.Conv2d(in_channels=128, out_channels=256, kernel_size=5, stride=2, padding=1) self.conv4 = nn.Conv2d(in_channels=256, out_channels=512, kernel_size=5, stride=2, padding=1) self.conv5 = nn.Conv2d(in_channels=512, out_channels=512, kernel_size=3, stride=2, padding=1) # Batch norm should be before relu self.bn1 = nn.BatchNorm2d(64) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(256) self.bn4 = nn.BatchNorm2d(512) self.bn5 = nn.BatchNorm2d(512) self.bn6 = nn.BatchNorm1d(4096) self.bn7 = nn.BatchNorm1d(1024) self.bn8 = nn.BatchNorm1d(512) self.dropout = nn.Dropout(p=0.8) self.fc1 = nn.Linear(18432, 4096) self.fc2 = nn.Linear(4096, 1024) self.fc3 = nn.Linear(1024, 512) self.fc_out_translation = nn.Linear(512, 3) self.fc_out_rotation = nn.Linear(512, 4) # instead of treating the relu as modules, we can treat them as functions. We can access them via torch funtional def forward(self, x, verbose=False): # this is where we pass the input into the module if verbose: print('shape ' + str(x.shape)) x = F.relu(self.bn1(self.conv1(x))) if verbose: print('layer1 shape ' + str(x.shape)) x = F.relu(self.bn2(self.conv2(x))) if verbose: print('layer2 shape ' + str(x.shape)) x = F.relu(self.bn3(self.conv3(x))) if verbose: print('layer3 shape ' + str(x.shape)) x = F.relu(self.bn4(self.conv4(x))) if verbose: print('layer4 shape ' + str(x.shape)) x = F.relu(self.bn5(self.conv5(x))) if verbose: print('layer5 shape ' + str(x.shape)) x = x.view(x.size(0), -1) if verbose: print('x shape ' + str(x.shape)) # x = F.dropout(x, p=0.5) # x = F.relu(self.fc1(x)) # if verbose: print('fc1 shape ' + str(x.shape)) # # x = F.relu(self.fc2(x)) # if verbose: print('fc2 shape ' + str(x.shape)) # # x = F.relu(self.fc3(x)) # if verbose: print('fc3 shape ' + str(x.shape)) x = F.relu(self.dropout(self.bn6(self.fc1(x)))) if verbose: print('fc1 shape ' + str(x.shape)) x = F.relu(self.bn7(self.fc2(x))) if verbose: print('fc2 shape ' + str(x.shape)) x = F.relu(self.bn8(self.fc3(x))) if verbose: print('fc3 shape ' + str(x.shape)) x_translation = self.fc_out_translation(x) if verbose: print('x_translation shape ' + str(x_translation.shape)) x_rotation = self.fc_out_rotation(x) if verbose: print('x_rotation shape ' + str(x_rotation.shape)) x_pose = torch.cat((x_translation, x_rotation), dim=1) return x_pose
45,692
44.784569
152
py
synfeal
synfeal-main/models/hourglass.py
import torch import torch.nn as nn import torch.nn.parallel import torch.utils.data import torch.nn.functional as F from torchvision import transforms, models # paper: https://arxiv.org/abs/1703.07971 # github: https://github.com/youngguncho/HourglassPose-Pytorch/blob/master/model.py class HourglassBatch(nn.Module): def __init__(self, pretrained, sum_mode=False, dropout_rate=0.5, aux_logits=False): super(HourglassBatch, self).__init__() self.sum_mode = sum_mode self.dropout_rate = dropout_rate self.aux_logits = aux_logits if pretrained: base_model = models.resnet34('ResNet34_Weights.DEFAULT') else: base_model = models.resnet34() # encoding blocks! self.init_block = nn.Sequential(*list(base_model.children())[:4]) self.res_block1 = base_model.layer1 self.res_block2 = base_model.layer2 self.res_block3 = base_model.layer3 self.res_block4 = base_model.layer4 # decoding blocks if sum_mode: self.deconv_block1 = nn.ConvTranspose2d(512, 256, kernel_size=( 3, 3), stride=(2, 2), padding=(1, 1), bias=False, output_padding=1) self.deconv_block2 = nn.ConvTranspose2d(256, 128, kernel_size=( 3, 3), stride=(2, 2), padding=(1, 1), bias=False, output_padding=1) self.deconv_block3 = nn.ConvTranspose2d(128, 64, kernel_size=( 3, 3), stride=(2, 2), padding=(1, 1), bias=False, output_padding=1) self.conv_block = nn.Conv2d(64, 32, kernel_size=( 3, 3), stride=(1, 1), padding=(1, 1), bias=False) else: # concatenation with the encoder feature vectors self.deconv_block1 = nn.ConvTranspose2d(512, 256, kernel_size=( 3, 3), stride=(2, 2), padding=(1, 1), bias=False, output_padding=1) self.deconv_block2 = nn.ConvTranspose2d(512, 128, kernel_size=( 3, 3), stride=(2, 2), padding=(1, 1), bias=False, output_padding=1) self.deconv_block3 = nn.ConvTranspose2d(256, 64, kernel_size=( 3, 3), stride=(2, 2), padding=(1, 1), bias=False, output_padding=1) self.conv_block = nn.Conv2d(128, 32, kernel_size=( 3, 3), stride=(1, 1), padding=(1, 1), bias=False) self.bn1 = nn.BatchNorm2d(256) self.bn2 = nn.BatchNorm2d(128) self.bn3 = nn.BatchNorm2d(64) self.bn4 = nn.BatchNorm2d(32) self.bn5 = nn.BatchNorm1d(1024) # Regressor self.fc_dim_reduce = nn.Linear(56 * 56 * 32, 1024) self.fc_trans = nn.Linear(1024, 3) self.fc_rot = nn.Linear(1024, 4) # Initialize Weights init_modules = [self.deconv_block1, self.deconv_block2, self.deconv_block3, self.conv_block, self.fc_dim_reduce, self.fc_trans, self.fc_rot] for module in init_modules: if isinstance(module, nn.ConvTranspose2d) or isinstance(module, nn.Linear) or isinstance(module, nn.Conv3d): nn.init.kaiming_normal_(module.weight) if module.bias is not None: nn.init.constant_(module.bias, 0) def forward(self, x): # conv x = self.init_block(x) x_res1 = self.res_block1(x) x_res2 = self.res_block2(x_res1) x_res3 = self.res_block3(x_res2) x_res4 = self.res_block4(x_res3) # Deconv x_deconv1 = self.bn1(F.relu(self.deconv_block1(x_res4))) if self.sum_mode: x_deconv1 = x_res3 + x_deconv1 else: x_deconv1 = torch.cat((x_res3, x_deconv1), dim=1) x_deconv2 = self.bn2(F.relu(self.deconv_block2(x_deconv1))) if self.sum_mode: x_deconv2 = x_res2 + x_deconv2 else: x_deconv2 = torch.cat((x_res2, x_deconv2), dim=1) x_deconv3 = self.bn3(F.relu(self.deconv_block3(x_deconv2))) if self.sum_mode: x_deconv3 = x_res1 + x_deconv3 else: x_deconv3 = torch.cat((x_res1, x_deconv3), dim=1) x_conv = self.bn4(F.relu(self.conv_block(x_deconv3))) x_linear = x_conv.view(x_conv.size(0), -1) x_linear = self.bn5(F.relu(self.fc_dim_reduce(x_linear))) x_linear = F.dropout(x_linear, p=self.dropout_rate, training=self.training) position = self.fc_trans(x_linear) rotation = self.fc_rot(x_linear) x_pose = torch.cat((position, rotation), dim=1) return x_pose
4,639
36.723577
120
py
synfeal
synfeal-main/process_dataset/src/validate_dataset.py
from utils import projectToCamera import numpy as np import os import shutil from dataset import Dataset from sensor_msgs.msg import PointField import sensor_msgs.point_cloud2 as pc2 # from utils import from utils_ros import read_pcd, write_pcd import random from os.path import exists import yaml from colorama import Fore import math import copy import cv2 class ValidateDataset(): def __init__(self): self.files = ['.pcd', '.rgb.png', '.pose.txt'] def resetConfig(self, config={}): self.config = config def duplicateDataset(self, dataset, suffix): # copy folder and create dataset object, return object path_dataset = dataset.path_seq path_validated_dataset = f'{dataset.path_seq}{suffix}' shutil.copytree(path_dataset, path_validated_dataset) return Dataset(path_seq=f'{dataset.seq}{suffix}') def numberOfPoints(self, dataset, frame = None): dct = {} if frame == None: # calculate number of points for all pointclouds for index in range(len(dataset)): n_points = read_pcd(f'{dataset.path_seq}/frame-{index:05d}.pcd').points dct[index] = n_points else: n_points = read_pcd(f'{dataset.path_seq}/frame-{frame:05d}.pcd').points dct[frame] = n_points return dct def numberOfNans(self, dataset, frame = None): dct = {} if frame == None: for index in range(len(dataset)): pc_raw = read_pcd(f'{dataset.path_seq}/frame-{index:05d}.pcd') pts = np.vstack([pc_raw.pc_data['x'], pc_raw.pc_data['y'], pc_raw.pc_data['z']]).T # stays NX3 dct[index] = np.sum(np.isnan(pts).any(axis=1)) else: pc_raw = read_pcd(f'{dataset.path_seq}/frame-{frame:05d}.pcd') pts = np.vstack([pc_raw.pc_data['x'], pc_raw.pc_data['y'], pc_raw.pc_data['z']]).T # stays NX3 dct[frame] = np.sum(np.isnan(pts).any(axis=1)) return dct def removeNansPointCloud(self, dataset, frame = None): fields = [PointField('x', 0, PointField.FLOAT32, 1), PointField('y', 4, PointField.FLOAT32, 1), PointField('z', 8, PointField.FLOAT32, 1), PointField('intensity', 12, PointField.FLOAT32, 1)] if frame == None: for index in range(len(dataset)): pc_raw = read_pcd(f'{dataset.path_seq}/frame-{index:05d}.pcd') pts = np.vstack([pc_raw.pc_data['x'], pc_raw.pc_data['y'], pc_raw.pc_data['z'], pc_raw.pc_data['intensity']]).T # stays NX4 pts = pts[~np.isnan(pts).any(axis=1)] # del os.remove(f'{dataset.path_seq}/frame-{index:05d}.pcd') # save new point clouds pc_msg =pc2.create_cloud(None, fields, pts) write_pcd(f'{dataset.path_seq}/frame-{index:05d}.pcd', pc_msg) else: pc_raw = read_pcd(f'{dataset.path_seq}/frame-{frame:05d}.pcd') pts = np.vstack([pc_raw.pc_data['x'], pc_raw.pc_data['y'], pc_raw.pc_data['z'], pc_raw.pc_data['intensity']]).T # stays NX4 pts = pts[~np.isnan(pts).any(axis=1)] # del os.remove(f'{dataset.path_seq}/frame-{frame:05d}.pcd') # save new point clouds pc_msg =pc2.create_cloud(None, fields, pts) write_pcd(f'{dataset.path_seq}/frame-{frame:05d}.pcd', pc_msg) def downsamplePointCloud(self, dataset, npoints): fields = [PointField('x', 0, PointField.FLOAT32, 1), PointField('y', 4, PointField.FLOAT32, 1), PointField('z', 8, PointField.FLOAT32, 1), PointField('intensity', 12, PointField.FLOAT32, 1)] for index in range(len(dataset)): pc_raw = read_pcd(f'{dataset.path_seq}/frame-{index:05d}.pcd') pts = np.vstack([pc_raw.pc_data['x'], pc_raw.pc_data['y'], pc_raw.pc_data['z'], pc_raw.pc_data['intensity']]).T # stays NX4 initial_npoints = pc_raw.points step = initial_npoints // npoints idxs = list(range(0,initial_npoints, step)) for i in range(len(idxs) - npoints): idxs.pop(random.randrange(len(idxs))) pts = pts[idxs,:] # del os.remove(f'{dataset.path_seq}/frame-{index:05d}.pcd') # save new point clouds pc_msg =pc2.create_cloud(None, fields, pts) write_pcd(f'{dataset.path_seq}/frame-{index:05d}.pcd', pc_msg) config = dataset.getConfig() config['npoints'] = npoints with open(f'{dataset.path_seq}/config.yaml', 'w') as file: yaml.dump(config, file) def scalePointCloud(self, dataset): fields = [PointField('x', 0, PointField.FLOAT32, 1), PointField('y', 4, PointField.FLOAT32, 1), PointField('z', 8, PointField.FLOAT32, 1), PointField('intensity', 12, PointField.FLOAT32, 1)] for index in range(len(dataset)): pc_raw = read_pcd(f'{dataset.path_seq}/frame-{index:05d}.pcd') pts = np.vstack([pc_raw.pc_data['x'], pc_raw.pc_data['y'], pc_raw.pc_data['z'], pc_raw.pc_data['intensity']]).T # stays NX4 pts = pts - np.expand_dims(np.mean(pts, axis=0), 0) # center dist = np.max(np.sqrt(np.sum(pts ** 2, axis=1)), 0) pts = pts / dist os.remove(f'{dataset.path_seq}/frame-{index:05d}.pcd') # save new point clouds pc_msg =pc2.create_cloud(None, fields, pts) write_pcd(f'{dataset.path_seq}/frame-{index:05d}.pcd', pc_msg) config = dataset.getConfig() config['scaled'] = True with open(f'{dataset.path_seq}/config.yaml', 'w') as file: yaml.dump(config, file) def invalidFrames(self, dataset): # return a list with invalid frames idxs = [] files = copy.deepcopy(self.files) config = dataset.getConfig() if config['fast']: files = ['.rgb.png','.pose.txt'] else: files = copy.deepcopy(self.files) files.append('.depth.png') for index in range(len(dataset)): for file in files: if not exists(f'{dataset.path_seq}/frame-{index:05d}{file}'): idxs.append(index) break return idxs def removeFrames(self, dataset, idxs): for idx in idxs: for file in os.listdir(f'{dataset.path_seq}'): if file.startswith(f'frame-{idx:05d}'): os.remove(f'{dataset.path_seq}/{file}') def reorganizeDataset(self, dataset): # here I assume the invalidFrames and removeFrames were called before. # last_pose_idx is the idx of the last frame. We cannot use len(dataset) because the dataset might be missing some frames! last_pose_idx = int(sorted([f for f in os.listdir(dataset.path_seq) if f.endswith('pose.txt')])[-1][6:11]) for idx in range(last_pose_idx+1): print(idx) if not exists(f'{dataset.path_seq}/frame-{idx:05d}.pose.txt'): # idx does not exists, so we have to rename the close one. print(f'{idx} is missing!!!') new_idx = None for idx2 in range(idx+1, last_pose_idx+1): print(f'trying {idx2}') if exists(f'{dataset.path_seq}/frame-{idx2:05d}.pose.txt'): new_idx = idx2 break if not new_idx==None: print(f'renaming idx {new_idx} to idx {idx}') for file in self.files: os.rename(f'{dataset.path_seq}/frame-{new_idx:05d}{file}', f'{dataset.path_seq}/frame-{idx:05d}{file}') else: print(f'No candidate to replace {idx}') def validateDataset(self, dataset): # update config, check if all point clouds have the same size, if any has nans # check for invalid frames idxs = self.invalidFrames(dataset) if idxs != []: print(f'{Fore.RED} There are invalid frames in the dataset! {Fore.RESET}') return False # # check for missing data # dct = self.numberOfNans(dataset) # n_nans = 0 # for count in dct.values(): # n_nans += count # if n_nans != 0: # print(f'{Fore.RED} There are Nans in the dataset! {Fore.RESET}') # return False # #check for point clouds of different size # dct = self.numberOfPoints(dataset) # number_of_points = list(dct.values()) # result = all(element == number_of_points[0] for element in number_of_points) # if not result: # print(f'{Fore.RED} Not all pointclouds have the same number of points! {Fore.RESET}') # return False config = dataset.getConfig() config['is_valid'] = True with open(f'{dataset.path_seq}/config.yaml', 'w') as file: yaml.dump(config, file) return True def mergeDatasets(self, dataset1, dataset2, dataset3_name): # both datasets must be valids # they should share the same number of points # if not (dataset1.getConfig()['is_valid'] and dataset2.getConfig()['is_valid']): # print(f'{Fore.RED} The datasets are not valid! Validate before merge. {Fore.RESET}') # return False # if not (dataset1.getConfig()['npoints'] == dataset2.getConfig()['npoints']): # print(f'{Fore.RED} The datasets dont have the same number of points! {Fore.RESET}') # return False # if not (dataset1.getConfig()['scaled'] == dataset2.getConfig()['scaled']): # print(f'{Fore.RED} Property scaled is different! {Fore.RESET}') # return False config = dataset1.getConfig() if config['fast']: files = ['.rgb.png','.pose.txt'] else: files = self.files size_dataset1 = len(dataset1) shutil.copytree(dataset1.path_seq, f'{dataset1.root}/{dataset3_name}') shutil.copytree(dataset2.path_seq, f'{dataset2.path_seq}_tmp') dataset3 = Dataset(path_seq=f'{dataset3_name}') dataset2_tmp = Dataset(path_seq=f'{dataset2.seq}_tmp') for idx in range(len(dataset2_tmp)): for file in files: os.rename(f'{dataset2_tmp.path_seq}/frame-{idx:05d}{file}', f'{dataset3.path_seq}/frame-{idx+size_dataset1:05d}{file}') shutil.rmtree(dataset2_tmp.path_seq) def createDepthImages(self, dataset, rescale): # loop through all point clouds config = dataset.getConfig() intrinsic = np.loadtxt(f'{dataset.path_seq}/depth_intrinsic.txt', delimiter=',') width = config['depth']['width'] height = config['depth']['height'] for idx in range(len(dataset)): pc_raw = read_pcd(f'{dataset.path_seq}/frame-{idx:05d}.pcd') pts = np.vstack([pc_raw.pc_data['x'], pc_raw.pc_data['y'], pc_raw.pc_data['z'], pc_raw.pc_data['intensity']]) # stays 4xN pixels, valid_pixels, dist = projectToCamera(intrinsic, [0, 0, 0, 0, 0], width, height, pts) range_sparse = np.zeros((height, width), dtype=np.float32) mask = 255 * np.ones((range_sparse.shape[0], range_sparse.shape[1]), dtype=np.uint8) for idx_point in range(0, pts.shape[1]): if valid_pixels[idx_point]: x0 = math.floor(pixels[0, idx_point]) y0 = math.floor(pixels[1, idx_point]) mask[y0, x0] = 0 range_sparse[y0, x0] = dist[idx_point] range_sparse = cv2.resize(range_sparse, (0, 0), fx=rescale, fy=rescale, interpolation=cv2.INTER_NEAREST) mask = cv2.resize(mask, (0, 0), fx=rescale, fy=rescale, interpolation=cv2.INTER_NEAREST) # Computing the dense depth map print('Computing inpaint ...') range_dense = cv2.inpaint(range_sparse, mask, 3, cv2.INPAINT_NS) print('Inpaint done') range_dense = cv2.resize(range_dense, (0, 0), fx=1 / rescale, fy=1 / rescale, interpolation=cv2.INTER_NEAREST) tmp = copy.deepcopy(range_dense) tmp = tmp * 1000.0 # to milimeters tmp = tmp.astype(np.uint16) cv2.imwrite(f'{dataset.path_seq}/frame-{idx:05d}.depth.png', tmp) print(f'Saved depth image {dataset.path_seq}/frame-{idx:05d}.depth.png') def createStatistics(self, dataset): # loop through all point clouds config = dataset.getConfig() config['statistics'] = {'B' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}, 'G' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}, 'R' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}, 'D' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}} for idx in range(len(dataset)): print(f'creating stats of frame {idx}') # Load RGB image cv_image = cv2.imread(f'{dataset.path_seq}/frame-{idx:05d}.rgb.png', cv2.IMREAD_UNCHANGED) #cv2.imshow('fig', cv_image) #cv2.waitKey(0) #print(cv_image.shape) blue_image = cv_image[:,:,0]/255 green_image = cv_image[:,:,1]/255 red_image = cv_image[:,:,2]/255 # cv2.imshow('fig', green_image) # cv2.waitKey(0) ## B channel config['statistics']['B']['max'][idx] = np.max(blue_image) config['statistics']['B']['min'][idx] = np.min(blue_image) config['statistics']['B']['mean'][idx] = np.mean(blue_image) config['statistics']['B']['std'][idx] = np.std(blue_image) ## G channel config['statistics']['G']['max'][idx] = np.max(green_image) config['statistics']['G']['min'][idx] = np.min(green_image) config['statistics']['G']['mean'][idx] = np.mean(green_image) config['statistics']['G']['std'][idx] = np.std(green_image) ## R channel config['statistics']['R']['max'][idx] = np.max(red_image) config['statistics']['R']['min'][idx] = np.min(red_image) config['statistics']['R']['mean'][idx] = np.mean(red_image) config['statistics']['R']['std'][idx] = np.std(red_image) # Load Depth image if not config['fast']: depth_image = cv2.imread(f'{dataset.path_seq}/frame-{idx:05d}.depth.png', cv2.IMREAD_UNCHANGED) depth_image = depth_image.astype(np.float32) / 1000.0 # to meters else: depth_image = -1 ## D channel config['statistics']['D']['max'][idx] = np.max(depth_image) config['statistics']['D']['min'][idx] = np.min(depth_image) config['statistics']['D']['mean'][idx] = np.mean(depth_image) config['statistics']['D']['std'][idx] = np.std(depth_image) config['statistics']['B']['max'] = round(float(np.mean(config['statistics']['B']['max'])),5) config['statistics']['B']['min'] = round(float(np.mean(config['statistics']['B']['min'])),5) config['statistics']['B']['mean'] = round(float(np.mean(config['statistics']['B']['mean'])),5) config['statistics']['B']['std'] = round(float(np.mean(config['statistics']['B']['std'])),5) config['statistics']['G']['max'] = round(float(np.mean(config['statistics']['G']['max'])),5) config['statistics']['G']['min'] = round(float(np.mean(config['statistics']['G']['min'])),5) config['statistics']['G']['mean'] = round(float(np.mean(config['statistics']['G']['mean'])),5) config['statistics']['G']['std'] = round(float(np.mean(config['statistics']['G']['std'])),5) config['statistics']['R']['max'] = round(float(np.mean(config['statistics']['R']['max'])),5) config['statistics']['R']['min'] = round(float(np.mean(config['statistics']['R']['min'])),5) config['statistics']['R']['mean'] = round(float(np.mean(config['statistics']['R']['mean'])),5) config['statistics']['R']['std'] = round(float(np.mean(config['statistics']['R']['std'])),5) config['statistics']['D']['max'] = round(float(np.mean(config['statistics']['D']['max'])),5) config['statistics']['D']['min'] = round(float(np.mean(config['statistics']['D']['min'])),5) config['statistics']['D']['mean'] = round(float(np.mean(config['statistics']['D']['mean'])),5) config['statistics']['D']['std'] = round(float(np.mean(config['statistics']['D']['std'])),5) dataset.setConfig(config) def createStatisticsRGB01(self, dataset): # loop through all point clouds config = dataset.getConfig() config['statistics'] = {'B' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}, 'G' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}, 'R' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}} for idx in range(len(dataset)): print(f'creating stats of frame {idx}') # Load RGB image cv_image = cv2.imread(f'{dataset.path_seq}/frame-{idx:05d}.rgb.png', cv2.IMREAD_UNCHANGED) #cv2.imshow('fig', cv_image) #cv2.waitKey(0) #print(cv_image.shape) blue_image = cv_image[:,:,0]/255 green_image = cv_image[:,:,1]/255 red_image = cv_image[:,:,2]/255 # cv2.imshow('fig', green_image) # cv2.waitKey(0) ## B channel config['statistics']['B']['max'][idx] = np.max(blue_image) config['statistics']['B']['min'][idx] = np.min(blue_image) config['statistics']['B']['mean'][idx] = np.mean(blue_image) config['statistics']['B']['std'][idx] = np.std(blue_image) ## G channel config['statistics']['G']['max'][idx] = np.max(green_image) config['statistics']['G']['min'][idx] = np.min(green_image) config['statistics']['G']['mean'][idx] = np.mean(green_image) config['statistics']['G']['std'][idx] = np.std(green_image) ## R channel config['statistics']['R']['max'][idx] = np.max(red_image) config['statistics']['R']['min'][idx] = np.min(red_image) config['statistics']['R']['mean'][idx] = np.mean(red_image) config['statistics']['R']['std'][idx] = np.std(red_image) config['statistics']['B']['max'] = round(float(np.mean(config['statistics']['B']['max'])),5) config['statistics']['B']['min'] = round(float(np.mean(config['statistics']['B']['min'])),5) config['statistics']['B']['mean'] = round(float(np.mean(config['statistics']['B']['mean'])),5) config['statistics']['B']['std'] = round(float(np.mean(config['statistics']['B']['std'])),5) config['statistics']['G']['max'] = round(float(np.mean(config['statistics']['G']['max'])),5) config['statistics']['G']['min'] = round(float(np.mean(config['statistics']['G']['min'])),5) config['statistics']['G']['mean'] = round(float(np.mean(config['statistics']['G']['mean'])),5) config['statistics']['G']['std'] = round(float(np.mean(config['statistics']['G']['std'])),5) config['statistics']['R']['max'] = round(float(np.mean(config['statistics']['R']['max'])),5) config['statistics']['R']['min'] = round(float(np.mean(config['statistics']['R']['min'])),5) config['statistics']['R']['mean'] = round(float(np.mean(config['statistics']['R']['mean'])),5) config['statistics']['R']['std'] = round(float(np.mean(config['statistics']['R']['std'])),5) dataset.setConfig(config) def processImages(self, dataset, technique, global_dataset): config = dataset.getConfig() config['statistics'] = {'B' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}, 'G' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}, 'R' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}, 'D' : {'max' : np.empty((len(dataset))), 'min' : np.empty((len(dataset))), 'mean' : np.empty((len(dataset))), 'std' : np.empty((len(dataset)))}} # Local Processing if global_dataset == None: for idx in range(len(dataset)): # RGB image bgr_image = cv2.imread(f'{dataset.path_seq}/frame-{idx:05d}.rgb.png', cv2.IMREAD_UNCHANGED) blue_image = bgr_image[:,:,0] green_image = bgr_image[:,:,1] red_image = bgr_image[:,:,2] depth_image = cv2.imread(f'{dataset.path_seq}/frame-{idx:05d}.depth.png', cv2.IMREAD_UNCHANGED) depth_image = depth_image.astype(np.float32) / 1000.0 # to meters if technique =='standardization': # B channel mean = np.mean(blue_image) std = np.std(blue_image) blue_image = (blue_image - mean) / std # G channel mean = np.mean(green_image) std = np.std(green_image) green_image = (green_image - mean) / std # R channel mean = np.mean(red_image) std = np.std(red_image) red_image = (red_image - mean) / std # D channel mean = np.mean(depth_image) std = np.std(depth_image) depth_image = (depth_image - mean) / std elif technique == 'normalization': # B channel min_v = np.min(blue_image) max_v = np.max(blue_image) blue_image = (blue_image - min_v) / (max_v - min_v) # G channel min_v = np.min(green_image) max_v = np.max(green_image) green_image = (green_image - min_v) / (max_v - min_v) # R channel min_v = np.min(red_image) max_v = np.max(red_image) red_image = (red_image - min_v) / (max_v - min_v) # D channel min_v = np.min(depth_image) max_v = np.max(depth_image) depth_image = (depth_image - min_v) / (max_v - min_v) else: print('Tehcnique not implemented. Available techniques are: normalization and standardization') exit(0) blue_image = blue_image.astype(np.float32) green_image = green_image.astype(np.float32) red_image = red_image.astype(np.float32) depth_image = depth_image.astype(np.float32) ## B channel config['statistics']['B']['max'][idx] = np.max(blue_image) config['statistics']['B']['min'][idx] = np.min(blue_image) config['statistics']['B']['mean'][idx] = np.mean(blue_image) config['statistics']['B']['std'][idx] = np.std(blue_image) ## G channel config['statistics']['G']['max'][idx] = np.max(green_image) config['statistics']['G']['min'][idx] = np.min(green_image) config['statistics']['G']['mean'][idx] = np.mean(green_image) config['statistics']['G']['std'][idx] = np.std(green_image) ## R channel config['statistics']['R']['max'][idx] = np.max(red_image) config['statistics']['R']['min'][idx] = np.min(red_image) config['statistics']['R']['mean'][idx] = np.mean(red_image) config['statistics']['R']['std'][idx] = np.std(red_image) ## D channel config['statistics']['D']['max'][idx] = np.max(depth_image) config['statistics']['D']['min'][idx] = np.min(depth_image) config['statistics']['D']['mean'][idx] = np.mean(depth_image) config['statistics']['D']['std'][idx] = np.std(depth_image) # joint BGR images as nparray bgr_image = cv2.merge([blue_image, green_image, red_image]) os.remove(f'{dataset.path_seq}/frame-{idx:05d}.depth.png') os.remove(f'{dataset.path_seq}/frame-{idx:05d}.rgb.png') np.save(f'{dataset.path_seq}/frame-{idx:05d}.depth.npy', depth_image) np.save(f'{dataset.path_seq}/frame-{idx:05d}.rgb.npy', bgr_image) #cv2.imwrite(f'{dataset.path_seq}/frame-{idx:05d}.depth.png', tmp) config['processing'] = {'global' : None, 'technique' : technique} config['statistics']['B']['max'] = round(float(np.mean(config['statistics']['B']['max'])),5) config['statistics']['B']['min'] = round(float(np.mean(config['statistics']['B']['min'])),5) config['statistics']['B']['mean'] = round(float(np.mean(config['statistics']['B']['mean'])),5) config['statistics']['B']['std'] = round(float(np.mean(config['statistics']['B']['std'])),5) config['statistics']['G']['max'] = round(float(np.mean(config['statistics']['G']['max'])),5) config['statistics']['G']['min'] = round(float(np.mean(config['statistics']['G']['min'])),5) config['statistics']['G']['mean'] = round(float(np.mean(config['statistics']['G']['mean'])),5) config['statistics']['G']['std'] = round(float(np.mean(config['statistics']['G']['std'])),5) config['statistics']['R']['max'] = round(float(np.mean(config['statistics']['R']['max'])),5) config['statistics']['R']['min'] = round(float(np.mean(config['statistics']['R']['min'])),5) config['statistics']['R']['mean'] = round(float(np.mean(config['statistics']['R']['mean'])),5) config['statistics']['R']['std'] = round(float(np.mean(config['statistics']['R']['std'])),5) config['statistics']['D']['max'] = round(float(np.mean(config['statistics']['D']['max'])),5) config['statistics']['D']['min'] = round(float(np.mean(config['statistics']['D']['min'])),5) config['statistics']['D']['mean'] = round(float(np.mean(config['statistics']['D']['mean'])),5) config['statistics']['D']['std'] = round(float(np.mean(config['statistics']['D']['std'])),5) dataset.setConfig(config) # Global Processing else: global_config = global_dataset.getConfig() global_stats = global_config['statistics'] for idx in range(len(dataset)): bgr_image = cv2.imread(f'{dataset.path_seq}/frame-{idx:05d}.rgb.png', cv2.IMREAD_UNCHANGED) blue_image = bgr_image[:,:,0] green_image = bgr_image[:,:,1] red_image = bgr_image[:,:,2] depth_image = cv2.imread(f'{dataset.path_seq}/frame-{idx:05d}.depth.png', cv2.IMREAD_UNCHANGED) depth_image = depth_image.astype(np.float32) / 1000.0 # to meters if technique =='standardization': blue_image = (blue_image - global_stats['B']['mean']) / global_stats['B']['std'] green_image = (green_image - global_stats['G']['mean']) / global_stats['G']['std'] red_image = (red_image - global_stats['R']['mean']) / global_stats['R']['std'] depth_image = (depth_image - global_stats['D']['mean']) / global_stats['D']['std'] elif technique == 'normalization': blue_image = (blue_image - global_stats['B']['min']) / (global_stats['B']['max'] - global_stats['B']['min']) green_image = (green_image - global_stats['G']['min']) / (global_stats['G']['max'] - global_stats['G']['min']) red_image = (red_image - global_stats['R']['min']) / (global_stats['R']['max'] - global_stats['R']['min']) depth_image = (depth_image - global_stats['D']['min']) / (global_stats['D']['max'] - global_stats['D']['min']) else: print('Tehcnique not implemented. Available techniques are: normalization and standardization') exit(0) blue_image = blue_image.astype(np.float32) green_image = green_image.astype(np.float32) red_image = red_image.astype(np.float32) depth_image = depth_image.astype(np.float32) ## B channel config['statistics']['B']['max'][idx] = np.max(blue_image) config['statistics']['B']['min'][idx] = np.min(blue_image) config['statistics']['B']['mean'][idx] = np.mean(blue_image) config['statistics']['B']['std'][idx] = np.std(blue_image) ## G channel config['statistics']['G']['max'][idx] = np.max(green_image) config['statistics']['G']['min'][idx] = np.min(green_image) config['statistics']['G']['mean'][idx] = np.mean(green_image) config['statistics']['G']['std'][idx] = np.std(green_image) ## R channel config['statistics']['R']['max'][idx] = np.max(red_image) config['statistics']['R']['min'][idx] = np.min(red_image) config['statistics']['R']['mean'][idx] = np.mean(red_image) config['statistics']['R']['std'][idx] = np.std(red_image) ## D channel config['statistics']['D']['max'][idx] = np.max(depth_image) config['statistics']['D']['min'][idx] = np.min(depth_image) config['statistics']['D']['mean'][idx] = np.mean(depth_image) config['statistics']['D']['std'][idx] = np.std(depth_image) # joint BGR images as nparray bgr_image = cv2.merge([blue_image, green_image, red_image]) os.remove(f'{dataset.path_seq}/frame-{idx:05d}.depth.png') os.remove(f'{dataset.path_seq}/frame-{idx:05d}.rgb.png') np.save(f'{dataset.path_seq}/frame-{idx:05d}.depth.npy', depth_image) np.save(f'{dataset.path_seq}/frame-{idx:05d}.rgb.npy', bgr_image) config['processing'] = {'global' : global_dataset.path_seq, 'technique' : technique} config['statistics']['B']['max'] = round(float(np.mean(config['statistics']['B']['max'])),5) config['statistics']['B']['min'] = round(float(np.mean(config['statistics']['B']['min'])),5) config['statistics']['B']['mean'] = round(float(np.mean(config['statistics']['B']['mean'])),5) config['statistics']['B']['std'] = round(float(np.mean(config['statistics']['B']['std'])),5) config['statistics']['G']['max'] = round(float(np.mean(config['statistics']['G']['max'])),5) config['statistics']['G']['min'] = round(float(np.mean(config['statistics']['G']['min'])),5) config['statistics']['G']['mean'] = round(float(np.mean(config['statistics']['G']['mean'])),5) config['statistics']['G']['std'] = round(float(np.mean(config['statistics']['G']['std'])),5) config['statistics']['R']['max'] = round(float(np.mean(config['statistics']['R']['max'])),5) config['statistics']['R']['min'] = round(float(np.mean(config['statistics']['R']['min'])),5) config['statistics']['R']['mean'] = round(float(np.mean(config['statistics']['R']['mean'])),5) config['statistics']['R']['std'] = round(float(np.mean(config['statistics']['R']['std'])),5) config['statistics']['D']['max'] = round(float(np.mean(config['statistics']['D']['max'])),5) config['statistics']['D']['min'] = round(float(np.mean(config['statistics']['D']['min'])),5) config['statistics']['D']['mean'] = round(float(np.mean(config['statistics']['D']['mean'])),5) config['statistics']['D']['std'] = round(float(np.mean(config['statistics']['D']['std'])),5) dataset.setConfig(config)
37,230
49.312162
140
py
synfeal
synfeal-main/process_dataset/src/__init__.py
0
0
0
py
synfeal
synfeal-main/process_dataset/scripts/reduce_dataset.py
#!/usr/bin/env python3 # stdlib import sys import argparse import copy # 3rd-party from dataset import Dataset import os import shutil import yaml def main(): parser = argparse.ArgumentParser(description='Validate dataset') parser.add_argument('-d', '--dataset', type=str, required=True, help='Name of the dataset') parser.add_argument('-dr', '--dataset_reduced', type=str, required=True, help='Suffix to append to the name of the dataset') parser.add_argument('-s', '--size', type=int, required=True, help='Sample size') arglist = [x for x in sys.argv[1:] if not x.startswith('__')] args = vars(parser.parse_args(args=arglist)) dataset = Dataset(path_seq=args['dataset']) path_root = dataset.root dataset_reduced_path = f'{path_root}/{args["dataset_reduced"]}' if os.path.exists(dataset_reduced_path): print(f'{dataset_reduced_path} already exits. Aborting reducing') exit(0) else: os.makedirs(dataset_reduced_path) # Create the new folder # get config config = dataset.getConfig() if 'statistics' in config: config.pop('statistics') if not config['fast']: files_to_copy = ['.pcd', '.rgb.png', '.depth.png','.pose.txt'] else: files_to_copy = ['.rgb.png','.pose.txt'] # copy intrinsics to both datasets for idx in range(len(dataset)): print(f'original idx: {idx}') if idx <= args['size']: print(f'copying {idx} to {idx} in {dataset_reduced_path}') for file in files_to_copy: shutil.copy2(f'{dataset.path_seq}/frame-{idx:05d}{file}', f'{dataset_reduced_path}/frame-{idx:05d}{file}') # copy intrinsics to both datasets shutil.copy2(f'{dataset.path_seq}/depth_intrinsic.txt', f'{dataset_reduced_path}/depth_intrinsic.txt') shutil.copy2(f'{dataset.path_seq}/rgb_intrinsic.txt', f'{dataset_reduced_path}/rgb_intrinsic.txt') config['raw'] = args['dataset_reduced'] with open(f'{dataset_reduced_path}/config.yaml', 'w') as f: yaml.dump(config, f) if __name__ == "__main__": main()
2,174
30.071429
128
py
synfeal
synfeal-main/synfeal_bringup/scripts/model_states_to_tf.py
#!/usr/bin/env python3 # -------------------------------------------------- # Adapted from http://wiki.ros.org/tf2/Tutorials/Writing%20a%20tf2%20broadcaster%20%28Python%29 # ------------------------------------------------- from functools import partial import rospy import tf2_ros import geometry_msgs.msg from gazebo_msgs.msg import ModelStates def callbackModelStatesReceived(msg, tf_broadcaster): childs = msg.name pose = msg.pose world = 'world' now = rospy.Time.now() # the gazebo has several models, so we have to pick the one we want if 'localbot' in childs: idx = childs.index('localbot') transform = geometry_msgs.msg.TransformStamped() transform.header.frame_id = world transform.child_frame_id = '/base_footprint' transform.header.stamp = now transform.transform.translation.x = pose[idx].position.x transform.transform.translation.y = pose[idx].position.y transform.transform.translation.z = pose[idx].position.z transform.transform.rotation.x = pose[idx].orientation.x transform.transform.rotation.y = pose[idx].orientation.y transform.transform.rotation.z = pose[idx].orientation.z transform.transform.rotation.w = pose[idx].orientation.w tf_broadcaster.sendTransform(transform) def main(): rospy.init_node('model_states_to_tf') rospy.Subscriber("/gazebo/model_states_throttle", ModelStates, partial(callbackModelStatesReceived, tf_broadcaster=tf2_ros.TransformBroadcaster())) rospy.spin() if __name__ == '__main__': main()
1,631
33.723404
105
py
synfeal
synfeal-main/synfeal_visualization/src/generate_real_vs_predicted.py
#!/usr/bin/env python3 import rospy import os from gazebo_msgs.srv import SetModelState, GetModelState, GetModelStateRequest, SetModelStateRequest from colorama import Fore from sensor_msgs.msg import Image from cv_bridge import CvBridge from utils import write_img class GenerateRealPredicted(): def __init__(self, model_name, results): self.set_state_service = rospy.ServiceProxy('/gazebo/set_model_state', SetModelState) self.model_name = model_name # model_name = 'localbot' self.bridge = CvBridge() self.folder = f'{results.path}/images' if not os.path.exists(self.folder): print(f'Creating folder {self.folder}') os.makedirs(self.folder) # Create the new folder else: print(f'{Fore.RED} {self.folder} already exists... Aborting GenerateRealPredicted initialization! {Fore.RESET}') exit(0) rospy.wait_for_service('/gazebo/get_model_state') self.get_model_state_service = rospy.ServiceProxy('/gazebo/get_model_state', GetModelState) def getPose(self): return self.get_model_state_service(self.model_name, 'world') def setPose(self, pose): req = SetModelStateRequest() # Create an object of type SetModelStateRequest req.model_state.model_name = self.model_name req.model_state.pose.position.x = pose.position.x req.model_state.pose.position.y = pose.position.y req.model_state.pose.position.z = pose.position.z req.model_state.pose.orientation.x = pose.orientation.x req.model_state.pose.orientation.y = pose.orientation.y req.model_state.pose.orientation.z = pose.orientation.z req.model_state.pose.orientation.w = pose.orientation.w req.model_state.reference_frame = 'world' self.set_state_service(req.model_state) def getImage(self): rgb_msg = rospy.wait_for_message('/kinect/rgb/image_raw', Image) return self.bridge.imgmsg_to_cv2(rgb_msg, "bgr8") # convert to opencv image def saveImage(self, filename, image): filename = f'{self.folder}/{filename}' write_img(filename, image)
2,221
39.4
124
py
synfeal
synfeal-main/produce_results/src/results.py
import numpy as np import pandas as pd import os import yaml from yaml.loader import SafeLoader from utils import matrixToXYZ, matrixToQuaternion, normalize_quat class Results(): def __init__(self, results_path): path=os.environ.get("SYNFEAL_DATASET") self.path = f'{path}/results/localbot/{results_path}' self.nframes = int(sum(f.endswith('.txt') for f in os.listdir(self.path))/2) self.csv = pd.read_csv(f'{self.path}/errors.csv') def __getitem__(self, index): # load pose matrix_predicted = np.loadtxt(f'{self.path}/frame-{index:05d}.predicted.pose.txt', delimiter=',') matrix_real = np.loadtxt(f'{self.path}/frame-{index:05d}.real.pose.txt', delimiter=',') quaternion_real = matrixToQuaternion(matrix_real) quaternion_real = normalize_quat(quaternion_real) xyz_real = matrixToXYZ(matrix_real) pose_real = np.append(xyz_real, quaternion_real) quaternion_predicted = matrixToQuaternion(matrix_predicted) quaternion_predicted = normalize_quat(quaternion_predicted) xyz_predicted = matrixToXYZ(matrix_predicted) pose_predicted = np.append(xyz_predicted, quaternion_predicted) return pose_real, pose_predicted def __len__(self): return self.nframes def getErrorsArrays(self): pos_error_array = self.csv.iloc[:-1]['position_error (m)'].to_numpy() rot_error_array = self.csv.iloc[:-1]['rotation_error (rads)'].to_numpy() return pos_error_array, rot_error_array def updateCSV(self): self.csv.to_csv(f'{self.path}/errors.csv', index=False, float_format='%.5f') def getConfig(self): with open(f'{self.path}/config.yaml') as f: config = yaml.load(f, Loader=SafeLoader) return config
1,911
32.54386
105
py
synfeal
synfeal-main/produce_results/src/save_results.py
#!/usr/bin/env python3 import os import yaml import pandas as pd import shutil import matplotlib.pyplot as plt from utils import write_transformation from colorama import Fore from datetime import datetime class SaveResults(): """ class to save results """ def __init__(self, output, model_path, seq_path, overwrite): # attribute initializer path=os.environ.get("SYNFEAL_DATASET") self.output_folder = f'{path}/results/localbot/{output}' self.model_path = model_path self.seq_path = seq_path if not os.path.exists(self.output_folder): print(f'Creating folder {self.output_folder}') os.makedirs(self.output_folder) # Create the new folder elif overwrite: print(f'Overwriting folder {self.output_folder}') shutil.rmtree(self.output_folder) os.makedirs(self.output_folder) # Create the new folder else: print(f'{Fore.RED} {self.output_folder} already exists... Aborting SaveResults initialization! {Fore.RESET}') exit(0) dt_now = datetime.now() # current date and time config = {'user' : os.environ["USER"], 'date' : dt_now.strftime("%d/%m/%Y, %H:%M:%S"), 'model_path' : self.model_path, 'seq_path' : self.seq_path} with open(f'{self.output_folder}/config.yaml', 'w') as file: yaml.dump(config, file) self.frame_idx = 0 # make sure to save as 00000 self.csv = pd.DataFrame(columns=('frame', 'position_error (m)', 'rotation_error (rads)')) print('SaveResults initialized properly') def saveTXT(self, real_transformation, predicted_transformation): filename = f'frame-{self.frame_idx:05d}' write_transformation(f'{self.output_folder}/{filename}.real.pose.txt', real_transformation) write_transformation(f'{self.output_folder}/{filename}.predicted.pose.txt', predicted_transformation) def updateCSV(self, position_error, rotation_error): row = {'frame' : f'{self.frame_idx:05d}', 'position_error (m)' : position_error, 'rotation_error (rads)' : rotation_error} self.csv = self.csv.append(row, ignore_index=True) def saveCSV(self): # save averages values in the last row mean_row = {'frame' : 'mean_values', 'position_error (m)' : self.csv.mean(axis=0).loc["position_error (m)"], 'rotation_error (rads)' : self.csv.mean(axis=0).loc["rotation_error (rads)"]} median_row = {'frame' : 'median_values', 'position_error (m)' : self.csv.median(axis=0).loc["position_error (m)"], 'rotation_error (rads)' : self.csv.median(axis=0).loc["rotation_error (rads)"]} self.csv = self.csv.append(mean_row, ignore_index=True) self.csv = self.csv.append(median_row, ignore_index=True) print(self.csv) self.csv.to_csv(f'{self.output_folder}/errors.csv', index=False, float_format='%.5f') def saveErrorsFig(self): frames_array = self.csv.iloc[:-2]['frame'].to_numpy().astype(int) pos_error_array = self.csv.iloc[:-2]['position_error (m)'].to_numpy() rot_error_array = self.csv.iloc[:-2]['rotation_error (rads)'].to_numpy() fig, (ax1, ax2) = plt.subplots(2, sharex=True) fig.suptitle('position and rotation errors') ax1.plot(frames_array, pos_error_array, 'cyan', label='position error') ax2.plot(frames_array, rot_error_array, 'navy', label='rotation error') ax2.set_xlabel('frame idx') ax2.set_ylabel('[rads]') ax1.set_ylabel('[m]') ax1.legend() ax2.legend() plt.savefig(f'{self.output_folder}/errors.png') def step(self): self.frame_idx+=1
4,142
37.719626
121
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/__init__.py
0
0
0
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/catalog/pylib_catalog.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 20 10:39:12 2021 @author: glavrent """ #load libraries #arithmetic libraries import numpy as np def IndexAvgColumns(df_data, col_idx, col2avg): ''' Average columns based on index column Parameters ---------- df_data : pd.dataframe Data data-frame. col_idx : str Name of index column. col2avg : list List of column names to be averaged. Returns ------- df_data : pd.dataframe Data data-frame. ''' #unique ids idx_array, inv_array = np.unique(df_data[col_idx], return_inverse=True) #iterate over columns for col in col2avg: #compute average values for all unique indices avg_vals = np.array([np.nanmean(df_data.loc[df_data[col_idx] == idx,col]) for idx in idx_array]) df_data.loc[:,col] = avg_vals[inv_array] return df_data def ColocatePt(df_flatfile, col_idx, col_coor, thres_dist=0.01, return_df_pt=False): ''' Colocate points (assign same ID) based on threshold distance. Parameters ---------- df_flatfile : pd.DataFrame Catalog flatfile. col_idx : str Name of index column. col_coor : list of str List of coordinate name columns. thres_dist : real, optional Value of threshold distance. The default is 0.01. return_df_pt : bool, optional Option for returning point data frame. The default is False. Returns ------- df_flatfile : pd.DataFrame Catalog flatfile with updated index column. df_pt: pd.DataFrame Point data frame with updated index column. ''' #dataframe with unique points _, pt_idx, pt_inv = np.unique(df_flatfile[col_idx], axis=0, return_index=True, return_inverse=True) df_pt = df_flatfile.loc[:,[col_idx] + col_coor].iloc[pt_idx,:] #find and merge collocated points for _, pt in df_pt.iterrows(): #distance between points dist2pt = np.linalg.norm((df_pt[col_coor] - pt[col_coor]).astype(float), axis=1) #indices of collocated points i_pt_coll = dist2pt < thres_dist #assign new id for collocated points df_pt.loc[i_pt_coll,col_idx] = pt[col_idx].astype(int) #update pt info to main catalog df_flatfile.loc[:,col_idx] = df_pt[col_idx].values[pt_inv] if not return_df_pt: return df_flatfile else: return df_flatfile, df_pt def UsableSta(mag_array, dist_array, df_coeffs): ''' Find records that meet the mag-distance limits Parameters ---------- mag_array : np.array Magnitude array. dist_array : np.array Distance array. df_coeffs : pd.DataFrame Coefficients dataframe. Returns ------- rec_lim : np.array logical array with True for records that meet M/R limits. ''' #rrup limit rrup_lim = dist_array <= df_coeffs.loc['max_rrup','coefficients'] #mag limit mag_min = (df_coeffs.loc['b1','coefficients'] + df_coeffs.loc['b1','coefficients'] * dist_array + df_coeffs.loc['b2','coefficients'] * dist_array**2) mag_lim = mag_array >= mag_min #find records that meet both conditions rec_lim = np.logical_and(rrup_lim, mag_lim) return rec_lim
3,361
26.557377
104
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/catalog/__init__.py
0
0
0
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/plotting/pylib_contour_plots.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sat Nov 9 13:12:38 2019 @author: glavrent """ ## load libraries #arithmetic import numpy as np from scipy.interpolate import griddata from scipy.ndimage import gaussian_filter #plotting import matplotlib import matplotlib.pyplot as plt from mpl_toolkits.axes_grid1 import make_axes_locatable #from matplotlib.ticker import FormatStrFormatter from matplotlib import ticker #base map from cartopy import config import cartopy.crs as ccrs import cartopy.feature as cfeature class FormatScalarFormatter(matplotlib.ticker.ScalarFormatter): def __init__(self, fformat="%1.1f", offset=True, mathText=True): self.fformat = fformat matplotlib.ticker.ScalarFormatter.__init__(self,useOffset=offset, useMathText=mathText) def _set_format(self, vmin, vmax): self.format = self.fformat if self._useMathText: self.format = '$%s$' % matplotlib.ticker._mathdefault(self.format) ## Main functions ##--------------------------------------- def PlotContourMapObs(cont_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', prj_map = False): ''' PlotContourMapObs: Input Arguments: cont_latlondata (np.array [n1,3]): contains the latitude, logitude and contour values cont_latlondata = [lat, long, data] cmin (double-opt): lower limit for color levels for contour plot cmax (double-opt): upper limit for color levels for contour plot title (str-opt): figure title cbar_label (str-opt): contour plot color bar label ptlevs (np.array-opt): color levels for points pt_label (str-opt): points color bar label log_cbar (bool-opt): if true use log-scale for contour plots frmt_clb string format color bar ticks Output Arguments: ''' plt_res = '50m' plt_scale = '50m' #number of interpolation points, x & y direction #ngridx = 5000 #ngridy = 5000 #ngridx = 500 #ngridy = 500 ngridx = 100 ngridy = 100 #create figure fig = plt.figure(figsize=(10, 10)) #fig = plt.figure(figsize=(15, 15)) #create basemap if prj_map == True: data_crs = ccrs.PlateCarree() ax = fig.add_subplot(1, 1, 1, projection=data_crs) else: data_crs = None ax = fig.add_subplot(1, 1, 1) #project contour data x_cont = cont_latlondata[:,1] y_cont = cont_latlondata[:,0] #interpolation grid x_int = np.linspace(x_cont.min(), x_cont.max(), ngridx) y_int = np.linspace(y_cont.min(), y_cont.max(), ngridy) X_grid, Y_grid = np.meshgrid(x_int, y_int) #interpolate contour data on grid if log_cbar: data_cont = np.log(cont_latlondata[:,2]) else: data_cont = cont_latlondata[:,2] data_grid = griddata((x_cont, y_cont) , data_cont, (X_grid, Y_grid), method='linear') #data colorbar cbmin = data_cont.min() if cmin is None else cmin cbmax = data_cont.max() if cmax is None else cmax clevs = np.linspace(cbmin, cbmax, 41).tolist() #plot interpolated data if prj_map == True: cs = ax.contourf(X_grid, Y_grid, data_grid, transform = data_crs, vmin=cmin, vmax=cmax, levels = clevs, zorder=3, alpha = 0.75) else: cs = ax.contourf(X_grid, Y_grid, data_grid, vmin=cmin, vmax=cmax, levels = clevs, zorder=3, alpha = 0.75) #color bar fmt_clb = ticker.FormatStrFormatter(frmt_clb) cbar_ticks = clevs[0:41:8] cbar = fig.colorbar(cs, boundaries=clevs, ticks=cbar_ticks, pad=0.05, orientation="horizontal", format=fmt_clb) # add colorbar if log_cbar: cbar_labels = [frmt_clb%np.exp(c_t) for c_t in cbar_ticks] cbar.set_ticklabels(cbar_labels) #add tick labs cbar.ax.tick_params(labelsize=18) if (not cbar_label is None): cbar.set_label(cbar_label, size=20) if prj_map == True: #add costal lines ax.coastlines(resolution=plt_res, edgecolor='black', zorder=5); #add state boundaries states = cfeature.NaturalEarthFeature(category='cultural', name='admin_1_states_provinces_lines', scale=plt_scale, facecolor='none') ax.add_feature(states, edgecolor='black', zorder=3) borders = cfeature.NaturalEarthFeature(category='cultural', name='admin_0_countries', scale=plt_scale, facecolor='none') ax.add_feature(borders, edgecolor='black', zorder=4) #add water bodies oceans = cfeature.NaturalEarthFeature(category='physical', name='ocean', facecolor='lightblue', scale=plt_scale) ax.add_feature(oceans, zorder=6) #add figure title if (not title is None): plt.title(title, fontsize=25) plt.xlabel('Latitude (deg)', fontsize=20) plt.ylabel('Longitude (deg)', fontsize=20) #grid lines if flag_grid: # gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True) gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True, linewidth=1, color='gray', alpha=0.5, linestyle='--') gl.xlabels_top = False gl.ylabels_right = False else: gl = None # fig.show() # fig.draw() fig.tight_layout() return fig, ax, cbar, data_crs, gl # Original PlotContourCAMap function #---- ---- ---- ---- ---- ---- ---- def PlotContourCAMapAdv(cont_latlondata, line_latlon=None, pt_latlondata=None, clevs=None, flag_grid=False, title=None, cbar_label=None, ptlevs = None, pt_label = None, log_cbar = False, frmt_clb = '%.2f', **kwargs): ''' PlotContourCAMapAdv: create a contour plot of the data in cont_latlondata Input Arguments: cont_latlondata (np.array [n1,3]): contains the latitude, logitude and contour values cont_latlondata = [lat, long, data] line_latlon (np.array-opt [n2,2]): contains the latitde and logitude coordinates of any lines pt_latlondata (np.array-opt [n3,(2,3)]): contains the latitude, logitude and values of disp points pt_latlondata = [lat, long, data-optional] clevs (np.array-opt): color levels for contour plot title (str-opt): figure title cbar_label (str-opt): contour plot color bar label ptlevs (np.array-opt): color levels for points pt_label (str-opt): points color bar label log_cbar (bool-opt): if true use log-scale for contour plots frmt_clb string format color bar ticks Output Arguments: ''' #additional input arguments flag_smooth = kwargs['flag_smooth'] if 'flag_smooth' in kwargs else False sig_smooth = kwargs['smooth_sig'] if 'smooth_sig' in kwargs else 0.1 plt_res = '10m' plt_scale = '10m' #number of interpolation points, x & y direction #ngridx = 5000 #ngridy = 5000 #ngridx = 500 #ngridy = 500 ngridx = 100 ngridy = 100 #create figure fig = plt.figure(figsize=(10, 10)) #fig = plt.figure(figsize=(15, 15)) #create basemap data_crs = ccrs.PlateCarree() ax = fig.add_subplot(1, 1, 1, projection=data_crs) #project contour data x_cont = cont_latlondata[:,1] y_cont = cont_latlondata[:,0] #interpolation grid x_int = np.linspace(x_cont.min(), x_cont.max(), ngridx) y_int = np.linspace(y_cont.min(), y_cont.max(), ngridy) X_grid, Y_grid = np.meshgrid(x_int, y_int) #interpolate contour data on grid data_cont = cont_latlondata[:,2] data_grid = griddata((x_cont, y_cont) , data_cont, (X_grid, Y_grid), method='linear') #smooth if flag_smooth: data_grid = gaussian_filter(data_grid, sigma=sig_smooth) #data colorbar if clevs is None: if not log_cbar: clevs = np.linspace(data_cont.min(),data_cont.max(),11).tolist() else: clevs = np.logspace(np.log10(data_cont.min()),np.log10(data_cont.max()),11).tolist() #plot interpolated data if not log_cbar: cs = ax.contourf(X_grid, Y_grid, data_grid, transform = data_crs, levels = clevs, zorder=3, alpha = 0.75) else: cs = ax.contourf(X_grid, Y_grid, data_grid, transform = data_crs, levels = clevs, zorder=3, alpha = 0.75, locator=ticker.LogLocator()) #color bar fmt_clb = ticker.FormatStrFormatter(frmt_clb) if not log_cbar: cbar = fig.colorbar(cs, boundaries = clevs, pad=0.05, orientation="horizontal", format=fmt_clb) # add colorbar else: cbar = fig.colorbar(cs, boundaries = clevs, pad=0.05, orientation="horizontal", format=fmt_clb) # add colorbar cbar.ax.tick_params(labelsize=18) if (not cbar_label is None): cbar.set_label(cbar_label, size=20) #plot line if not line_latlon is None: ax.plot(line_latlon[:,1], line_latlon[:,0], latlon = True, linewidth=3, color='k', zorder= 5 ) #plot points if not pt_latlondata is None: if np.size(pt_latlondata,1) == 2: ax.plot(pt_latlondata[:,1], pt_latlondata[:,0], 'o', latlon=True, color = 'k', markersize = 4, zorder = 8) elif np.size(pt_latlondata,1) == 2: raise ValueError('Unimplemented plotting option') #add costal lines ax.coastlines(resolution=plt_res, edgecolor='black', zorder=5); #add state boundaries states = cfeature.NaturalEarthFeature(category='cultural', name='admin_1_states_provinces_lines', scale=plt_scale, facecolor='none') ax.add_feature(states, edgecolor='black', zorder=3) ax.add_feature(cfeature.BORDERS, zorder=4) #add oceans oceans = cfeature.NaturalEarthFeature(category='physical', name='ocean', facecolor='lightblue', scale=plt_scale) ax.add_feature(oceans, zorder=6) #add figure title if (not title is None): plt.title(title, fontsize=25) plt.xlabel('Latitude (deg)', fontsize=20) plt.ylabel('Longitude (deg)', fontsize=20) #grid lines if flag_grid: # gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True) gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True, linewidth=1, color='gray', alpha=0.5, linestyle='--') gl.xlabels_top = False gl.ylabels_right = False else: gl = None fig.tight_layout() return fig, ax, cbar, data_crs, gl # Updated PlotContourCAMap function #---- ---- ---- ---- ---- ---- ---- def PlotContourCAMap(cont_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', cmap = 'viridis', **kwargs): ''' PlotContourCAMap: simplifed function to create a contour plot of the data in cont_latlondata Input Arguments: cont_latlondata (np.array [n1,3]): contains the latitude, logitude and contour values cont_latlondata = [lat, long, data] cmin (double-opt): lower limit for color levels for contour plot cmax (double-opt): upper limit for color levels for contour plot title (str-opt): figure title cbar_label (str-opt): contour plot color bar label ptlevs (np.array-opt): color levels for points pt_label (str-opt): points color bar label log_cbar (bool-opt): if true use log-scale for contour plots frmt_clb string format color bar ticks Output Arguments: ''' #additional input arguments flag_smooth = kwargs['flag_smooth'] if 'flag_smooth' in kwargs else False sig_smooth = kwargs['smooth_sig'] if 'smooth_sig' in kwargs else 0.1 intrp_method = kwargs['intrp_method'] if 'intrp_method' in kwargs else 'linear' plt_res = '50m' plt_scale = '50m' #number of interpolation points, x & y direction #ngridx = 5000 #ngridy = 5000 ngridx = 500 ngridy = 500 #ngridx = 100 #ngridy = 100 #create figure fig = plt.figure(figsize=(10, 10)) #fig = plt.figure(figsize=(15, 15)) #create basemap data_crs = ccrs.PlateCarree() ax = fig.add_subplot(1, 1, 1, projection=data_crs) #project contour data x_cont = cont_latlondata[:,1] y_cont = cont_latlondata[:,0] #interpolation grid x_int = np.linspace(x_cont.min(), x_cont.max(), ngridx) y_int = np.linspace(y_cont.min(), y_cont.max(), ngridy) X_grid, Y_grid = np.meshgrid(x_int, y_int) #interpolate contour data on grid if log_cbar: data_cont = np.log(cont_latlondata[:,2]) else: data_cont = cont_latlondata[:,2] data_grid = griddata((x_cont, y_cont) , data_cont, (X_grid, Y_grid), method=intrp_method ) #smooth if flag_smooth: data_grid = gaussian_filter(data_grid, sigma=sig_smooth) #data colorbar cbmin = data_cont.min() if cmin is None else cmin cbmax = data_cont.max() if cmax is None else cmax clevs = np.linspace(cbmin, cbmax, 41).tolist() #plot interpolated data cs = ax.contourf(X_grid, Y_grid, data_grid, transform = data_crs, vmin=cmin, vmax=cmax, levels = clevs, zorder=3, alpha = 0.75, cmap=cmap) #color bar #import pdb; pdb.set_trace() fmt_clb = ticker.FormatStrFormatter(frmt_clb) cbar_ticks = clevs[0:41:10] cbar = fig.colorbar(cs, boundaries=clevs, ticks=cbar_ticks, pad=0.05, orientation="horizontal", format=fmt_clb) # add colorbar if log_cbar: cbar_labels = [frmt_clb%np.exp(c_t) for c_t in cbar_ticks] cbar.set_ticklabels(cbar_labels) cbar.ax.tick_params(labelsize=18) if (not cbar_label is None): cbar.set_label(cbar_label, size=20) #add costal lines ax.coastlines(resolution=plt_res, edgecolor='black', zorder=5); #add state boundaries states = cfeature.NaturalEarthFeature(category='cultural', name='admin_1_states_provinces_lines', scale=plt_scale, facecolor='none') ax.add_feature(states, edgecolor='black', zorder=3) borders = cfeature.NaturalEarthFeature(category='cultural', name='admin_0_countries', scale=plt_scale, facecolor='none') ax.add_feature(borders, edgecolor='black', zorder=4) #add oceans oceans = cfeature.NaturalEarthFeature(category='physical', name='ocean', scale=plt_scale) ax.add_feature(oceans, zorder=6) #add figure title if (not title is None): plt.title(title, fontsize=25) plt.xlabel('Latitude (deg)', fontsize=20) plt.ylabel('Longitude (deg)', fontsize=20) #grid lines if flag_grid: # gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True) gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True, linewidth=1, color='gray', alpha=0.5, linestyle='--') gl.xlabels_top = False gl.ylabels_right = False else: gl = None # fig.show() # fig.draw() fig.tight_layout() return fig, ax, cbar, data_crs, gl # PlotContourSloveniaMap function #---- ---- ---- ---- ---- ---- ---- def PlotContourSloveniaMap(cont_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', **kwargs): ''' PlotContourCAMap: simplifed create a contour plot of the data in cont_latlondata Input Arguments: cont_latlondata (np.array [n1,3]): contains the latitude, logitude and contour values cont_latlondata = [lat, long, data] cmin (double-opt): lower limit for color levels for contour plot cmax (double-opt): upper limit for color levels for contour plot title (str-opt): figure title cbar_label (str-opt): contour plot color bar label ptlevs (np.array-opt): color levels for points pt_label (str-opt): points color bar label log_cbar (bool-opt): if true use log-scale for contour plots frmt_clb string format color bar ticks Output Arguments: ''' plt_res = '50m' plt_scale = '50m' #number of interpolation points, x & y direction #ngridx = 5000 #ngridy = 5000 #ngridx = 500 #ngridy = 500 ngridx = 100 ngridy = 100 #create figure fig = plt.figure(figsize=(10, 10)) #fig = plt.figure(figsize=(15, 15)) #create basemap data_crs = ccrs.PlateCarree() ax = fig.add_subplot(1, 1, 1, projection=data_crs) #project contour data x_cont = cont_latlondata[:,1] y_cont = cont_latlondata[:,0] #interpolation grid x_int = np.linspace(x_cont.min(), x_cont.max(), ngridx) y_int = np.linspace(y_cont.min(), y_cont.max(), ngridy) X_grid, Y_grid = np.meshgrid(x_int, y_int) #interpolate contour data on grid if log_cbar: data_cont = np.log(cont_latlondata[:,2]) else: data_cont = cont_latlondata[:,2] data_grid = griddata((x_cont, y_cont) , data_cont, (X_grid, Y_grid), method='linear') #smooth if (kwargs['flag_smooth'] if 'flag_smooth' in kwargs else False): sig_smooth = kwargs['smooth_sig'] if 'smooth_sig' in kwargs else 0.1 data_grid = gaussian_filter(data_grid, sigma=sig_smooth) #data colorbar cbmin = data_cont.min() if cmin is None else cmin cbmax = data_cont.max() if cmax is None else cmax clevs = np.linspace(cbmin, cbmax, 41).tolist() #plot interpolated data cs = ax.contourf(X_grid, Y_grid, data_grid, transform = data_crs, vmin=cmin, vmax=cmax, levels = clevs, zorder=3, alpha = 0.75) #color bar fmt_clb = ticker.FormatStrFormatter(frmt_clb) cbar_ticks = clevs[0:41:8] cbar = fig.colorbar(cs, boundaries=clevs, ticks=cbar_ticks, pad=0.05, orientation="horizontal", format=fmt_clb) # add colorbar if log_cbar: cbar_labels = [frmt_clb%np.exp(c_t) for c_t in cbar_ticks] cbar.set_ticklabels(cbar_labels) cbar.ax.tick_params(labelsize=18) if (not cbar_label is None): cbar.set_label(cbar_label, size=20) #add costal lines ax.coastlines(resolution=plt_res, edgecolor='black', zorder=5); #add state boundaries #states = cfeature.NaturalEarthFeature(category='cultural', name='admin_1_states_provinces_lines', # scale=plt_scale, facecolor='none') #ax.add_feature(states, edgecolor='black', zorder=3) borders = cfeature.NaturalEarthFeature(category='cultural', name='admin_0_countries', scale=plt_scale, facecolor='none') ax.add_feature(borders, edgecolor='black', zorder=4) #ax.add_feature(cfeature.BORDERS, zorder=4) #add oceans oceans = cfeature.NaturalEarthFeature(category='physical', name='ocean', facecolor='lightblue', scale=plt_scale) ax.add_feature(oceans, zorder=6) #add figure title if (not title is None): plt.title(title, fontsize=25) plt.xlabel('Latitude (deg)', fontsize=20) plt.ylabel('Longitude (deg)', fontsize=20) #grid lines if flag_grid: # gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True) gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True, linewidth=1, color='gray', alpha=0.5, linestyle='--') gl.xlabels_top = False gl.ylabels_right = False else: gl = None # fig.show() # fig.draw() fig.tight_layout() return fig, ax, cbar, data_crs, gl # Scatter plot function #---- ---- ---- ---- ---- ---- ---- def PlotScatterCAMap(scat_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', alpha_v = 0.7, cmap='seismic', marker_size=10.): ''' PlotContourCAMap: create a contour plot of the data in cont_latlondata Input Arguments: scat_latlondata (np.array [n1,(3,4)]): contains the latitude, logitude, contour values, and size values (optional) scat_latlondata = [lat, long, data_color, data_size] cmin (double-opt): lower limit for color levels for contour plot cmax (double-opt): upper limit for color levels for contour plot title (str-opt): figure title cbar_label (str-opt): contour plot color bar label ptlevs (np.array-opt): color levels for points pt_label (str-opt): points color bar label log_cbar (bool-opt): if true use log-scale for contour plots frmt_clb: string format color bar ticks alpha_v: opacity value cmap: color palette marker_size: marker size, if scat_latlondata dimensions is [n1, 3] Output Arguments: ''' #import pdb; pdb.set_trace() plt_res = '10m' plt_scale = '10m' #create figure fig = plt.figure(figsize=(10, 10)) #fig = plt.figure(figsize=(15, 15)) #create basemap data_crs = ccrs.PlateCarree() ax = fig.add_subplot(1, 1, 1, projection=data_crs) #project contour data x_scat = scat_latlondata[:,1] y_scat = scat_latlondata[:,0] #color scale if log_cbar: data_scat_c = np.log(scat_latlondata[:,2]) else: data_scat_c = scat_latlondata[:,2] #size scale if scat_latlondata.shape[1] > 3: data_scat_s = scat_latlondata[:,3] else: data_scat_s = marker_size * np.ones(data_scat_c.shape) #data colorbar cbmin = data_scat_c.min() if cmin is None else cmin cbmax = data_scat_c.max() if cmax is None else cmax clevs = np.linspace(cbmin, cbmax, 41).tolist() #plot scatter bubble plot data cs = ax.scatter(x_scat, y_scat, s = data_scat_s, c = data_scat_c, transform = data_crs, vmin=cmin, vmax=cmax, zorder=3, alpha=alpha_v, cmap=cmap) #color bar #import pdb; pdb.set_trace() fmt_clb = ticker.FormatStrFormatter(frmt_clb) cbar_ticks = clevs[0:41:8] cbar = fig.colorbar(cs, boundaries=clevs, ticks=cbar_ticks, pad=0.05, orientation="horizontal", format=fmt_clb) # add colorbar if log_cbar: cbar_labels = [frmt_clb%np.exp(c_t) for c_t in cbar_ticks] cbar.set_ticklabels(cbar_labels) cbar.ax.tick_params(labelsize=18) if (not cbar_label is None): cbar.set_label(cbar_label, size=20) #add costal lines ax.coastlines(resolution=plt_res, edgecolor='black', zorder=5); #add state boundaries states = cfeature.NaturalEarthFeature(category='cultural', name='admin_1_states_provinces_lines', scale=plt_scale, facecolor='none') ax.add_feature(states, edgecolor='black', zorder=3) ax.add_feature(cfeature.BORDERS, zorder=4) #oceans oceans = cfeature.NaturalEarthFeature(category='physical', name='ocean', facecolor='lightblue', scale=plt_scale) ax.add_feature(oceans, zorder=2) #add figure title if (not title is None): plt.title(title, fontsize=25) plt.xlabel('Latitude (deg)', fontsize=20) plt.ylabel('Longitude (deg)', fontsize=20) #grid lines if flag_grid: # gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True) gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True, linewidth=1, color='gray', alpha=0.5, linestyle='--') gl.xlabels_top = False gl.ylabels_right = False else: gl = None # fig.show() # fig.draw() fig.tight_layout() return fig, ax, cbar, data_crs, gl # Updated PlotContourCAMap function #---- ---- ---- ---- ---- ---- ---- def PlotCellsCAMap(cell_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', alpha_v = .8, cell_size = 50, cmap='seismic'): ''' PlotCellsCAMap: PlotCellsCAMap function to create a contour plot of the data in cont_latlondata Input Arguments: cell_latlondata (np.array [n1,3]): contains the latitude, logitude and color values cell_latlondata = [lat, long, data] cmin (double-opt): lower limit for color levels for contour plot cmax (double-opt): upper limit for color levels for contour plot title (str-opt): figure title cbar_label (str-opt): contour plot color bar label ptlevs (np.array-opt): color levels for points pt_label (str-opt): points color bar label log_cbar (bool-opt): if true use log-scale for contour plots frmt_clb string format color bar ticks Output Arguments: ''' plt_res = '50m' plt_scale = '50m' #create figure fig = plt.figure(figsize=(10, 10)) #create basemap data_crs = ccrs.PlateCarree() ax = fig.add_subplot(1, 1, 1, projection=data_crs) #project contour data x_cell = cell_latlondata[:,1] y_cell = cell_latlondata[:,0] #contour transfomration if log_cbar: data_cell = np.log(cell_latlondata[:,2]) else: data_cell = cell_latlondata[:,2] #data colorbar cbmin = data_cell.min() if cmin is None else cmin cbmax = data_cell.max() if cmax is None else cmax clevs = np.linspace(cbmin, cbmax, 41).tolist() #plot interpolated data cs = ax.scatter(x_cell, y_cell, s = cell_size, c = data_cell, transform = data_crs, vmin=cmin, vmax=cmax, zorder=3, alpha = alpha_v, cmap=cmap) #cs = ax.contourf(X_grid, Y_grid, data_grid, transform = data_crs, vmin=cmin, vmax=cmax, levels = clevs, zorder=3, alpha = 0.75) #color bar #import pdb; pdb.set_trace() fmt_clb = ticker.FormatStrFormatter(frmt_clb) cbar_ticks = clevs[0:41:8] cbar = fig.colorbar(cs, boundaries=clevs, ticks=cbar_ticks, pad=0.05, orientation="horizontal", format=fmt_clb) # add colorbar if log_cbar: cbar_labels = [frmt_clb%np.exp(c_t) for c_t in cbar_ticks] cbar.set_ticklabels(cbar_labels) cbar.ax.tick_params(labelsize=18) if (not cbar_label is None): cbar.set_label(cbar_label, size=20) #add costal lines ax.coastlines(resolution=plt_res, edgecolor='black', zorder=5); #add state boundaries states = cfeature.NaturalEarthFeature(category='cultural', name='admin_1_states_provinces_lines', scale=plt_scale, facecolor='none') ax.add_feature(states, edgecolor='black', zorder=3) borders = cfeature.NaturalEarthFeature(category='cultural', name='admin_0_countries', scale=plt_scale, facecolor='none') ax.add_feature(borders, edgecolor='black', zorder=4) #add oceans #ax.stock_img() oceans = cfeature.NaturalEarthFeature(category='physical', name='ocean', facecolor='lightblue', scale=plt_scale) ax.add_feature(oceans, zorder=2) #add figure title if (not title is None): ax.set_title(title, fontsize=25) ax.set_xlabel('Latitude (deg)', fontsize=20) ax.set_ylabel('Longitude (deg)', fontsize=20) #grid lines if flag_grid: # gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True) gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True, linewidth=1, color='gray', alpha=0.5, linestyle='--') gl.xlabels_top = False gl.ylabels_right = False else: gl = None # fig.show() # fig.draw() fig.tight_layout() return fig, ax, cbar, data_crs, gl # Plotting coefficient function #---- ---- ---- ---- ---- ---- ---- #plotting of median values coefficients def PlotCoeffCAMapMed(cont_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', **kwargs): cmap = 'seismic' fig, ax, cbar, data_crs, gl = PlotContourCAMap(cont_latlondata, cmin=cmin, cmax=cmax, flag_grid=flag_grid, title=title, cbar_label=cbar_label, log_cbar = log_cbar, frmt_clb = frmt_clb, cmap = cmap, **kwargs) return fig, ax, cbar, data_crs, gl #plotting of epistemic uncertainty coefficients def PlotCoeffCAMapSig(cont_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', **kwargs): cmap = 'Purples_r' fig, ax, cbar, data_crs, gl = PlotContourCAMap(cont_latlondata, cmin=cmin, cmax=cmax, flag_grid=flag_grid, title=title, cbar_label=cbar_label, log_cbar = log_cbar, frmt_clb = frmt_clb, cmap = cmap, **kwargs) return fig, ax, cbar, data_crs, gl #plotting of median values of cells def PlotCellsCAMapMed(cell_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', alpha_v = .8, cell_size = 50): cmap = 'seismic' fig, ax, cbar, data_crs, gl = PlotCellsCAMap(cell_latlondata, cmin=cmin, cmax=cmax, flag_grid=flag_grid, title=title, cbar_label=cbar_label, log_cbar=log_cbar, frmt_clb=frmt_clb, alpha_v=alpha_v, cell_size=cell_size, cmap=cmap) return fig, ax, cbar, data_crs, gl #plotting of mono-color increasing values of cells def PlotCellsCAMapInc(cell_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', alpha_v = .8, cell_size = 50): cmap = 'Reds' fig, ax, cbar, data_crs, gl = PlotCellsCAMap(cell_latlondata, cmin=cmin, cmax=cmax, flag_grid=flag_grid, title=title, cbar_label=cbar_label, log_cbar=log_cbar, frmt_clb=frmt_clb, alpha_v=alpha_v, cell_size=cell_size, cmap=cmap) return fig, ax, cbar, data_crs, gl #plotting of epistemic uncertainty of cells def PlotCellsCAMapSig(cell_latlondata, cmin=None, cmax=None, flag_grid=False, title=None, cbar_label=None, log_cbar = False, frmt_clb = '%.2f', alpha_v = .8, cell_size = 50): cmap = 'Purples_r' fig, ax, cbar, data_crs, gl = PlotCellsCAMap(cell_latlondata, cmin=cmin, cmax=cmax, flag_grid=flag_grid, title=title, cbar_label=cbar_label, log_cbar=log_cbar, frmt_clb=frmt_clb, alpha_v=alpha_v, cell_size=cell_size, cmap=cmap ) return fig, ax, cbar, data_crs, gl # Base plot function #---- ---- ---- ---- ---- ---- ---- def PlotMap(lat_lims = None, lon_lims = None, flag_grid=False, title=None): ''' PlotContourCAMap: simplifed function to create a contour plot of the data in cont_latlondata Input Arguments: line_latlondata (np.array [n1,3]): contains the latitude, logitude and contour values cont_latlondata = [lat, long, data] cmin (double-opt): lower limit for color levels for contour plot cmax (double-opt): upper limit for color levels for contour plot title (str-opt): figure title cbar_label (str-opt): contour plot color bar label ptlevs (np.array-opt): color levels for points pt_label (str-opt): points color bar label log_cbar (bool-opt): if true use log-scale for contour plots frmt_clb string format color bar ticks Output Arguments: ''' plt_res = '50m' plt_scale = '50m' #create figure fig = plt.figure(figsize=(10, 10)) #fig = plt.figure(figsize=(15, 15)) #create basemap data_crs = ccrs.PlateCarree() ax = fig.add_subplot(1, 1, 1, projection=data_crs) if lat_lims: ax.set_xlim(lon_lims) if lon_lims: ax.set_ylim(lat_lims) #add land zones lands = cfeature.LAND ax.add_feature(lands, zorder=1) #add costal lines ax.coastlines(resolution=plt_res, edgecolor='black', zorder=3); #add state boundaries states = cfeature.NaturalEarthFeature(category='cultural', name='admin_1_states_provinces_lines', scale=plt_scale, facecolor='none') ax.add_feature(states, edgecolor='black', zorder=4) borders = cfeature.NaturalEarthFeature(category='cultural', name='admin_0_countries', scale=plt_scale, facecolor='none') ax.add_feature(borders, edgecolor='black', zorder=5) #add oceans oceans = cfeature.NaturalEarthFeature(category='physical', name='ocean', facecolor='lightblue', scale=plt_scale) ax.add_feature(oceans, zorder=2) #add figure title if (not title is None): plt.title(title, fontsize=25) plt.xlabel('Latitude (deg)', fontsize=20) plt.ylabel('Longitude (deg)', fontsize=20) #grid lines if flag_grid: # gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True) gl = ax.gridlines(crs=ccrs.PlateCarree(), draw_labels=True, linewidth=1, color='gray', alpha=0.5, linestyle='--') gl.xlabels_top = False gl.ylabels_right = False else: gl = None # fig.show() # fig.draw() # fig.tight_layout() return fig, ax, data_crs, gl
35,537
40.613583
154
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/plotting/__init__.py
0
0
0
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/ground_motions/pylib_gmm_eas.py
# ba18.py # Conversion of Jeff Bayless' MATLAB code to Python # Including class ba18 # I've tried to avoid mixed UPPER and lower case variable names # e.g. Mbreak, Rrup, Vsref #arithmetic libraries import numpy as np import numpy.matlib from scipy import linalg as scipylalg from scipy import sparse as scipysp #geographic coordinates import pyproj #statistics libraries import pandas as pd #geometric libraries from shapely.geometry import Point as shp_pt, Polygon as shp_poly def SlicingSparceMat(mat_sp, i_rows, j_col): '''Slice sparse matrix''' return np.array([mat_sp.getcol(i_r).toarray().flatten()[j_col] for i_r in i_rows]) def QuartCos(per, x0, x, flag_left = False): y = np.cos( 2.*np.pi*(x-x0)/per ) if flag_left: y[np.logical_or(x < x0-per/4, x > x0)] = 0. else: y[np.logical_or(x < x0, x > x0+per/4)] = 0. return y def QuadCosTapper(freq, freq_nerg): #boxcar at intermediate frequencies i_box = np.logical_and(freq >= freq_nerg.min(), freq <= freq_nerg.max()) y_box = np.zeros(len(freq)) y_box[i_box] = 1. #quarter cosine left taper per = 2 * freq_nerg.min() y_tpl = QuartCos(per, freq_nerg.min(), freq, flag_left=True) #quarter cosine right taper per = 2 * freq_nerg.max() y_tpr = QuartCos(per, freq_nerg.max(), freq) #combined tapering function y_tapper = np.array([y_box, y_tpl, y_tpr]).max(axis=0) return y_tapper def TriagTapper(freq, freq_nerg): fn_min = freq_nerg.min() fn_max = freq_nerg.max() #triangular window f_win = np.array([0.5*fn_min, fn_min, fn_max, 1.5*fn_max]) y_win = np.array([0., 1., 1., 0.]) #triangular tapering function y_tapper = np.interp(np.log(freq), np.log(f_win), y_win) return y_tapper def ConvertPandasDf2NpArray(df_array): array = df_array.values if isinstance(df_array, pd.DataFrame) or isinstance(df_array, pd.Series) else df_array return array class BA18: def __init__(self, file=None): ''' Constructor for this class Read CSV file of BA18 coefficients, frequency range: 0.1 - 100 Hz Parameters ---------- file : string, optional file name for coefficients. The default is None. ''' if file is None: file = '/mnt/halcloud_nfs/glavrent/Research/Nonerg_CA_GMM/Analyses/Python_lib/ground_motions/Bayless_ModelCoefs.csv' df = pd.read_csv(file, index_col=0) df = df.head(301) # Frequencies 0.1 - 24 Hz self.freq = df.index.values # Median FAS parameters self.b1 = df.c1.values self.b2 = df.c2.values self.b3quantity = df['(c2-c3)/cn'].values self.b3 = df.c3.values self.bn = df.cn.values self.bm = df.cM .values self.b4 = df.c4.values self.b5 = df.c5.values self.b6 = df.c6.values self.bhm = df.chm.values self.b7 = df.c7.values self.b8 = df.c8.values self.b9 = df.c9.values self.b10 = df.c10.values self.b11a = df.c11a.values self.b11b = df.c11b.values self.b11c = df.c11c.values self.b11d = df.c11d.values self.b1a = df.c1a.values self.b1a[239:] = 0 # Non-linear site parameters self.f3 = df.f3.values self.f4 = df.f4.values self.f5 = df.f5.values # Aleatory variability parameters self.s1 = df.s1.values self.s2 = df.s2.values self.s3 = df.s3.values self.s4 = df.s4.values self.s5 = df.s5.values self.s6 = df.s6.values # Constants self.b4a = -0.5 self.vsref = 1000 self.mbreak = 6.0 #bedrock anelastic attenuation self.b7rock = self.b7.copy() #frequency limits # self.maxfreq = 23.988321 self.maxfreq = self.freq.max() self.minfreq = self.freq.min() def EasBase(self, mag, rrup, vs30, ztor, fnorm, z1, regid, flag_keep_b7 = True): # note Z1 must be provided in km z1ref = (1/1000) * np.exp(-7.67/4 * np.log((vs30**4+610**4)/(1360**4+610**4)) ) if vs30<=200: self.b11 = self.b11a if vs30>200 and vs30<=300: self.b11 = self.b11b if vs30>300 and vs30<=500: self.b11 = self.b11c if vs30>500: self.b11 = self.b11d if z1 is None or np.isnan(z1): z1 = self.Z1(vs30, regid=1) # Compute lnFAS by summing contributions, including linear site response lnfas = self.b1 + self.b2*(mag-self.mbreak) lnfas += self.b3quantity*np.log(1+np.exp(self.bn*(self.bm-mag))) lnfas += self.b4*np.log(rrup+self.b5*np.cosh(self.b6*np.maximum(mag-self.bhm,0))) lnfas += (self.b4a-self.b4) * np.log( np.sqrt(rrup**2+50**2) ) lnfas += self.b7 * rrup if flag_keep_b7 else 0. lnfas += self.b8 * np.log( min(vs30,1000) / self.vsref ) lnfas += self.b9 * min(ztor,20) lnfas += self.b10 * fnorm lnfas += self.b11 * np.log( (min(z1,2) + 0.01) / (z1ref + 0.01) ) # this is the linear spectrum up to maxfreq=23.988321 Hz maxfreq = 23.988321 imax = np.where(self.freq==maxfreq)[0][0] fas_lin = np.exp(lnfas) # Extrapolate to 100 Hz fas_maxfreq = fas_lin[imax] # Kappa kappa = np.exp(-0.4*np.log(vs30/760)-3.5) # Diminuition operator D = np.exp(-np.pi*kappa*(self.freq[imax:] - maxfreq)) fas_lin = np.append(fas_lin[:imax], fas_maxfreq * D) # Compute non-linear site response # get the EAS_rock at 5 Hz (no c8, c11 terms) vref=760 #row = df.iloc[df.index == 5.011872] i5 = np.where(self.freq==5.011872) lnfasrock5Hz = self.b1[i5] lnfasrock5Hz += self.b2[i5]*(mag-self.mbreak) lnfasrock5Hz += self.b3quantity[i5]*np.log(1+np.exp(self.bn[i5]*(self.bm[i5]-mag))) lnfasrock5Hz += self.b4[i5]*np.log(rrup+self.b5[i5]*np.cosh(self.b6[i5]*max(mag-self.bhm[i5],0))) lnfasrock5Hz += (self.b4a-self.b4[i5])*np.log(np.sqrt(rrup**2+50**2)) lnfasrock5Hz += self.b7rock[i5]*rrup lnfasrock5Hz += self.b9[i5]*min(ztor,20) lnfasrock5Hz += self.b10[i5]*fnorm # Compute PGA_rock extimate from 5 Hz FAS IR = np.exp(1.238+0.846*lnfasrock5Hz) # apply the modified Hashash model self.f2 = self.f4*( np.exp(self.f5*(min(vs30,vref)-360)) - np.exp(self.f5*(vref-360)) ) fnl0 = self.f2 * np.log((IR+self.f3)/self.f3) fnl0[np.where(fnl0==min(fnl0))[0][0]:] = min(fnl0) fas_nlin = np.exp( np.log(fas_lin) + fnl0 ) # Aleatory variability if mag<4: tau = self.s1 phi_s2s = self.s3 phi_ss = self.s5 if mag>6: tau = self.s2 phi_s2s = self.s4 phi_ss = self.s6 if mag >= 4 and mag <= 6: tau = self.s1 + ((self.s2-self.s1)/2)*(mag-4) phi_s2s = self.s3 + ((self.s4-self.s3)/2)*(mag-4) phi_ss = self.s5 + ((self.s6-self.s5)/2)*(mag-4) sigma = np.sqrt(tau**2 + phi_s2s**2 + phi_ss**2 + self.b1a**2); return self.freq, fas_nlin, fas_lin, sigma def EasBaseArray(self, mag, rrup, vs30, ztor, fnorm, z1=None, regid=1, flag_keep_b7=True): #convert eq parameters to np.arrays mag = np.array([mag]).flatten() rrup = np.array([rrup]).flatten() vs30 = np.array([vs30]).flatten() ztor = np.array([ztor]).flatten() fnorm = np.array([fnorm]).flatten() z1 = np.array([self.Z1(vs, regid) for vs in vs30]) if z1 is None else np.array([z1]).flatten() #number of scenarios npt = len(mag) #input assertions assert( np.all(npt == np.array([len(rrup),len(vs30),len(ztor),len(fnorm),len(z1)])) ),'Error. Inconsistent number of gmm parameters' #compute fas for all scenarios fas_nlin = list() fas_lin = list() sigma = list() for k, (m, r, vs, zt, fn, z_1) in enumerate(zip(mag, rrup, vs30, ztor, fnorm, z1)): ba18_base = self.EasBase(m, r, vs, zt, fn, z_1, regid, flag_keep_b7)[1:] fas_nlin.append(ba18_base[0]) fas_lin.append(ba18_base[1]) sigma.append(ba18_base[2]) #combine them to np.arrays fas_nlin = np.vstack(fas_nlin) fas_lin = np.vstack(fas_lin) sigma = np.vstack(sigma) # if npt == 1 and flag_flatten: # fas_nlin = fas_nlin.flatten() # fas_lin = fas_lin.flatten() # sigma = sigma.flatten() #return self.EasBase(mag, rrup, vs30, ztor, fnorm, z1, regid, flag_keep_b7) return self.freq, fas_nlin, fas_lin, sigma def Eas(self, mag, rrup, vs30, ztor, fnorm, z1=None, regid=1, flag_keep_b7=True, flag_flatten=True): ''' Computes BA18 EAS GMM for all frequencies Parameters ---------- mag : real moment magnitude [3-8]. rrup : real Rupture distance in kilometers (km) [0-300]. vs30 : real site-specific Vs30 = slowness-averaged shear wavespeed of upper 30 m (m/s) [120-1500]. ztor : real depth to top of rupture (km) [0-20]. fnorm : real 1 for normal faults and 0 for all other faulting types (no units) [0 or 1]. z1 : real, optional site-specific depth to shear wavespeed of 1 km/s (km) [0-2]. The default is =None. regid : int, optional DESCRIPTION. The default is =1. Returns ------- freq : np.array frequency array. fas_nlin : np.array fas array with nonlinear site response. fas_lin : np.array fas array with linear site response. sigma : np.array standard deviation array. ''' #return self.EasBase(mag, rrup, vs30, ztor, fnorm, z1, regid, flag_keep_b7) # return self.EasBaseArray(mag, rrup, vs30, ztor, fnorm, z1, regid, flag_keep_b7, flag_flatten) freq, fas_nlin, fas_lin, sigma = self.EasBaseArray(mag, rrup, vs30, ztor, fnorm, z1, regid, flag_keep_b7) #flatten arrays if only one datapoint if fas_nlin.shape[0] == 1 and flag_flatten: fas_nlin = fas_nlin.flatten() fas_lin = fas_lin.flatten() sigma = sigma.flatten() return freq, fas_nlin, fas_lin, sigma def EasF(self, freq, mag, rrup, vs30, ztor, fnorm, z1=None, regid=1, flag_keep_b7 = True, flag_flatten=True): ''' Computes BA18 EAS GMM for frequency of interest Parameters ---------- mag : real moment magnitude [3-8]. rrup : real Rupture distance in kilometers (km) [0-300]. vs30 : real site-specific Vs30 = slowness-averaged shear wavespeed of upper 30 m (m/s) [120-1500]. ztor : real depth to top of rupture (km) [0-20]. fnorm : real 1 for normal faults and 0 for all other faulting types (no units) [0 or 1]. z1 : real, optional site-specific depth to shear wavespeed of 1 km/s (km) [0-2]. The default is =None. regid : int, optional DESCRIPTION. The default is =1. Returns ------- freq : real frequency of interest. fas_nlin : real fas with nonlinear site response for frequency of interest. fas_lin : real fas with linear site response for frequency of interest. sigma : real standard deviation of frequency of interest. ''' #convert freq to numpy array freq = np.array([freq]).flatten() #frequency tolerance f_tol = 1e-4 #compute fas for all frequencies freq_all, fas_all, fas_lin_all, sig_all = self.EasBaseArray(mag, rrup, vs30, ztor, fnorm, z1, regid, flag_keep_b7) #find eas for frequency of interest if np.all([np.isclose(f, freq_all, f_tol).any() for f in freq]): # i_f = np.array([np.where(np.isclose(f, freq_all, f_tol))[0] for f in freq]).flatten() i_f = np.array([np.argmin(np.abs(f-freq_all)) for f in freq]).flatten() freq = freq_all[i_f] fas = fas_all[:,i_f] fas_lin = fas_lin_all[:,i_f] sigma = sig_all[:,i_f] else: fas = np.vstack([np.exp(np.interp(np.log(np.abs(freq)), np.log(freq_all), np.log(fas), left=-np.nan, right=-np.nan)) for fas in fas_all]) fas_lin = np.vstack([np.exp(np.interp(np.log(np.abs(freq)), np.log(freq_all), np.log(fas_l), left=-np.nan, right=-np.nan)) for fas_l in fas_lin_all]) sigma = np.vstack([ np.interp(np.log(np.abs(freq)), np.log(freq_all), sig, left=-np.nan, right=-np.nan) for sig in sig_all]) #if one scenario flatten arrays if fas.shape[0] == 1 and flag_flatten: fas = fas.flatten() fas_lin = fas_lin.flatten() sigma = sigma.flatten() return fas, fas_lin, sigma def GetFreq(self): return np.array(self.freq) def Z1(self, vs30, regid=1): ''' Compute Z1.0 based on Vs30 for CA and JP Parameters ---------- vs30 : real Time average shear-wave velocity. regid : int, optional Region ID. The default is 1. Returns ------- real Depth to a shear wave velocity of 1000m/sec. ''' if regid == 1: #CA z_1 = -7.67/4. * np.log((vs30**4+610.**4)/(1360.**4+610.**4)) elif regid == 10: #JP z_1 = -5.23/4. * np.log((vs30**4+412.**4)/(1360.**4+412.**4)) return 1/1000*np.exp(z_1)
14,209
36.005208
161
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/ground_motions/pylib_Willis15CA_Vs30.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Feb 2 19:01:47 2021 @author: glavrent """ #load variables import pathlib import numpy as np import rasterio class Willis15Vs30CA: def __init__(self, fname_vs30map_med=None, fname_vs30map_sig=None): #file path root = pathlib.Path(__file__).parent #vs30 data filenames fname_vs30map_med = '/mnt/halcloud_nfs/glavrent/Research/Other_projects/VS30_CA/data/California_vs30_Wills15_hybrid_7p5c.tif' if fname_vs30map_med is None else fname_vs30map_med fname_vs30map_sig = '/mnt/halcloud_nfs/glavrent/Research/Other_projects/VS30_CA/data/California_vs30_Wills15_hybrid_7p5c_sd.tif' if fname_vs30map_sig is None else fname_vs30map_sig #load vs30 data # self.vs30map_med = rasterio.open(root / 'data/California_vs30_Wills15_hybrid_7p5c.tif') # self.vs30map_sig = rasterio.open(root / 'data/California_vs30_Wills15_hybrid_7p5c_sd.tif') self.vs30map_med = rasterio.open( fname_vs30map_med ) self.vs30map_sig = rasterio.open( fname_vs30map_sig ) def lookup(self, lonlats): return ( np.fromiter(self.vs30map_med.sample(lonlats, 1), np.float), np.fromiter(self.vs30map_sig.sample(lonlats, 1), np.float) ) def test_lookup(self): medians, stds = list(self.lookup([(-122.258, 37.875), (-122.295, 37.895)])) np.testing.assert_allclose(medians, [733.4, 351.9], rtol=0.01) np.testing.assert_allclose(stds, [0.432, 0.219], rtol=0.01)
1,564
36.261905
188
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/ground_motions/__init__.py
0
0
0
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/ground_motions/pylib_NGMM_prediction.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sat Aug 20 14:54:54 2022 @author: glavrent """ # Packages # --------------------------- #arithmetic libraries import numpy as np from scipy import linalg as scipylinalg from sklearn.gaussian_process.kernels import Matern #user functions import pylib_kernels as pylib_kern import pylib_cell_dist as pylib_cells # Non-ergodic GMM effects prediction # --------------------------- def PredictNErgEffects(n_samp, nerg_coeff_info, df_scen_predict, df_nerg_coeffs, nerg_catten_info=None, df_cell_info=None, df_nerg_cellatten=None): ''' Predict non-egodic ground motion effects Parameters ---------- n_samp : int Number of samples. nerg_coeff_info : dict Non-ergodic coefficient information dictionary. df_scen_predict : pd.dataframe Prediction scenarios. df_nerg_coeffs : pd.dataframe Regressed non-ergodic coefficients . nerg_catten_info : dict, optional cell-specific anelastic attenuation information dictionary. The default is None. df_cell_info : pd.dataframe, optional Cell info dataframe. The default is None. df_nerg_cellatten : pd.dataframe, optional Regressed anelastic attenuation coefficients. The default is None. Returns ------- nerg_effects_prdct_samp : np.array Samples of total non-ergodic effects. nerg_vcm_prdct_samp : TYPE Samples of spatially varying component of non-ergodic effects. nerg_atten_prdct_samp : TYPE Samples of anelastic attenuation component of non-ergodic effects. nerg_effects_prdct_mu : TYPE Mean of total non-ergodic effects. nerg_effects_prdct_sig : TYPE Standard deviation of total non-ergodic effects. nerg_vcm_cmp : list List with individual components of spatially varying non-ergodic effects. nerg_atten_cmp : list List with individual components of anelast attenuation. ''' #number of prediction scenarios n_predict = len(df_scen_predict) # VCM component # --- --- --- --- #initialize vcm samples nerg_vcm_prdct_samp = np.zeros(shape=(n_predict,n_samp)) nerg_vcm_prdct_mu = np.zeros(shape=n_predict) nerg_vcm_prdct_var = np.zeros(shape=n_predict) nerg_vcm_cmp = {} #iterate over non-ergodic coefficients for nerg_c in nerg_coeff_info: #kernel type k_type = nerg_coeff_info[nerg_c]['kernel_type'] #hyper-parameters if 'hyp' in nerg_coeff_info[nerg_c]: hyp_param = nerg_coeff_info[nerg_c]['hyp'] hyp_mean_c = hyp_param['mean_c'] if (('mean_c' in hyp_param) and (not hyp_param['mean_c'] is None)) else 0 hyp_ell = hyp_param['ell'] if (('ell' in hyp_param) and (not hyp_param['ell'] is None)) else 0 hyp_omega = hyp_param['omega'] if (('omega' in hyp_param) and (not hyp_param['omega'] is None)) else 0 hyp_pi = hyp_param['pi'] if (('pi' in hyp_param) and (not hyp_param['pi'] is None)) else 0 hyp_nu = hyp_param['nu'] if (('nu' in hyp_param) and (not hyp_param['nu'] is None)) else 0 #mean and std of non-ergodic coefficients at known locations c_mean_train = df_nerg_coeffs.loc[:,nerg_coeff_info[nerg_c]['coeff'][0]].values c_sig_train = df_nerg_coeffs.loc[:,nerg_coeff_info[nerg_c]['coeff'][1]].values #non-ergodic coefficient scaling c_scl = np.ones(n_predict) if nerg_coeff_info[nerg_c]['scaling'] is None else df_nerg_coeffs.loc[:,nerg_coeff_info[nerg_c]['scaling']].values if k_type == 0: #constan assert(len(np.unique(c_mean_train))==1) #mean and std of non-ergodic coefficient c_mean_train = c_mean_train[0] c_sig_train = c_sig_train[0] #draw random samples c_prdct_samp = np.random.normal(loc=c_mean_train, scale=c_sig_train, size=n_samp) #sample non-ergodic coefficient for prediction scenarios c_prdct_samp = np.full((n_predict,n_samp), c_prdct_samp) #mean and sigma c_prdct_mu = np.full(n_predict, c_mean_train) c_prdct_sig = np.full(n_predict, c_sig_train) if k_type == 1: #group #group ids in training data id_train = df_nerg_coeffs.loc[:,nerg_coeff_info[nerg_c]['cor_info']].values id_train, idx_train = np.unique(id_train, axis=0, return_index=True) #group ids in prediction data id_prdct = df_scen_predict.loc[:,nerg_coeff_info[nerg_c]['cor_info']].values id_prdct, inv_prdct = np.unique(id_prdct, axis=0, return_inverse=True) #mean and std of non-ergodic coefficient c_mean_train = c_mean_train[idx_train] c_sig_train = c_sig_train[idx_train] #compute mean and cov of non-erg coeffs for prediction scenarios c_prdct_mu, _, c_prdct_cov = pylib_kern.PredictGroupKern(id_prdct, id_train, c_train_mu=c_mean_train, c_train_sig=c_sig_train, hyp_mean_c=hyp_mean_c, hyp_omega=hyp_omega) #sample non-ergodic coefficient for prediction scenarios c_prdct_samp = MVNRnd(mean=c_prdct_mu, cov=c_prdct_cov, n_samp=n_samp) c_prdct_samp = c_prdct_samp[inv_prdct,:] #mean and sigma c_prdct_mu = c_prdct_mu[inv_prdct] c_prdct_sig = np.sqrt( np.diag(c_prdct_cov) )[inv_prdct] if k_type == 2: #exponetial #coordinates of training data t_train = df_nerg_coeffs.loc[:,nerg_coeff_info[nerg_c]['cor_info']].values t_train, idx_train = np.unique(t_train, axis=0, return_index=True) #coordinates of prediction data t_prdct = df_scen_predict.loc[:,nerg_coeff_info[nerg_c]['cor_info']].values t_prdct, inv_prdct = np.unique(t_prdct, axis=0, return_inverse=True) #mean and std of non-ergodic coefficient c_mean_train = c_mean_train[idx_train] c_sig_train = c_sig_train[idx_train] #compute mean and cov of non-erg coeffs for prediction scenarios c_prdct_mu, _, c_prdct_cov = pylib_kern.PredictExpKern(t_prdct, t_train, c_train_mu=c_mean_train, c_train_sig=c_sig_train, hyp_mean_c=hyp_mean_c, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi) #sample non-ergodic coefficient for prediction scenarios c_prdct_samp = MVNRnd(mean=c_prdct_mu, cov=c_prdct_cov, n_samp=n_samp) c_prdct_samp = c_prdct_samp[inv_prdct,:] #mean and sigma c_prdct_mu = c_prdct_mu[inv_prdct] c_prdct_sig = np.sqrt( np.diag(c_prdct_cov) )[inv_prdct] if k_type == 3: #squared exponetial #coordinates of training data t_train = df_nerg_coeffs.loc[:,nerg_coeff_info[nerg_c]['cor_info']].values t_train, idx_train = np.unique(t_train, axis=0, return_index=True) #coordinates of prediction data t_prdct = df_scen_predict.loc[:,nerg_coeff_info[nerg_c]['cor_info']].values t_prdct, inv_prdct = np.unique(t_prdct, axis=0, return_inverse=True) #mean and std of non-ergodic coefficient c_mean_train = c_mean_train[idx_train] c_sig_train = c_sig_train[idx_train] #compute mean and cov of non-erg coeffs for prediction scenarios c_prdct_mu, _, c_prdct_cov = pylib_kern.PredictSqExpKern(t_prdct, t_train, c_train_mu=c_mean_train, c_train_sig=c_sig_train, hyp_mean_c=hyp_mean_c, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi) #sample non-ergodic coefficient for prediction scenarios c_prdct_samp = MVNRnd(mean=c_prdct_mu, cov=c_prdct_cov, n_samp=n_samp) c_prdct_samp = c_prdct_samp[inv_prdct,:] #mean and sigma c_prdct_mu = c_prdct_mu[inv_prdct] c_prdct_sig = np.sqrt( np.diag(c_prdct_cov) )[inv_prdct] if k_type == 4: #Matern kernel function #coordinates of training data t_train = df_nerg_coeffs.loc[:,nerg_coeff_info[nerg_c]['cor_info']].values t_train, idx_train = np.unique(t_train, axis=0, return_index=True) #coordinates of prediction data t_prdct = df_scen_predict.loc[:,nerg_coeff_info[nerg_c]['cor_info']].values t_prdct, inv_prdct = np.unique(t_prdct, axis=0, return_inverse=True) #mean and std of non-ergodic coefficient c_mean_train = c_mean_train[idx_train] c_sig_train = c_sig_train[idx_train] #compute mean and cov of non-erg coeffs for prediction scenarios c_prdct_mu, _, c_prdct_cov = pylib_kern.PredictMaternKern(t_prdct, t_train, c_train_mu=c_mean_train, c_train_sig=c_sig_train, hyp_mean_c=hyp_mean_c, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, hyp_nu=hyp_nu) #sample non-ergodic coefficient for prediction scenarios c_prdct_samp = MVNRnd(mean=c_prdct_mu, cov=c_prdct_cov, n_samp=n_samp) c_prdct_samp = c_prdct_samp[inv_prdct,:] #mean and sigma c_prdct_mu = c_prdct_mu[inv_prdct] c_prdct_sig = np.sqrt( np.diag(c_prdct_cov) )[inv_prdct] #add contribution of non-ergodic effect nerg_vcm_prdct_samp += c_scl[:,np.newaxis] * c_prdct_samp #mean and std contribution of non-ergodic effect nerg_vcm_prdct_mu += c_scl * c_prdct_mu nerg_vcm_prdct_var += c_scl**2 * c_prdct_sig**2 #summarize individual components nerg_vcm_cmp[nerg_c] = [c_scl * c_prdct_mu, c_scl * c_prdct_sig, c_scl[:,np.newaxis] * c_prdct_samp] # Anelastic attenuation # --- --- --- --- #initialize anelastic attenuation nerg_atten_prdct_samp = np.zeros(shape=(n_predict,n_samp)) nerg_atten_prdct_mu = np.zeros(shape=n_predict) nerg_atten_prdct_var = np.zeros(shape=n_predict) nerg_atten_cmp = {} if not nerg_catten_info is None: #cell edge coordinates for path seg calculation ct4dist = df_cell_info.loc[:,['q1X', 'q1Y', 'q1Z', 'q8X', 'q8Y', 'q8Z']].values #cell limts c_lmax = ct4dist[:,[3,4,5]].max(axis=0) c_lmin = ct4dist[:,[0,1,2]].min(axis=0) #compute cell-path cell_path = np.zeros([n_predict, len(df_cell_info)]) for j, (rsn, scn_p) in enumerate(df_scen_predict.iterrows()): pt1 = scn_p[['eqX','eqY','eqZ']].values.astype(float) pt2 = np.hstack([scn_p[['staX','staY']].values, 0]).astype(float) #check limits assert(np.logical_and(pt1>=c_lmin, pt1<=c_lmax).all()),'Error. Eq outside cell domain for rsn: %i'%rsn assert(np.logical_and(pt2>=c_lmin, pt2<=c_lmax).all()),'Error. Sta outside cell domain for rsn: %i'%rsn #cell paths for pt1 - pt2 cell_path[j,:] = pylib_cells.ComputeDistGridCells(pt1,pt2,ct4dist, flagUTM=True) #keep only cells with non-zero paths ca_valid = cell_path.sum(axis=0) > 0 cell_path = cell_path[:,ca_valid] df_cell_info = df_cell_info.loc[ca_valid,:] #iterate over anelastic attenuation components for nerg_ca in nerg_catten_info: #kernel type k_type = nerg_catten_info[nerg_ca]['kernel_type'] #mean and std anelastic attenuation cells ca_mean_train = df_nerg_cellatten.loc[:,nerg_catten_info[nerg_ca]['catten'][0]].values ca_sig_train = df_nerg_cellatten.loc[:,nerg_catten_info[nerg_ca]['catten'][1]].values #hyper-parameters hyp_param = nerg_catten_info[nerg_ca]['hyp'] hyp_mean_ca = hyp_param['mean_ca'] if (('mean_ca' in hyp_param) and (not hyp_param['mean_ca'] is None)) else 0 hyp_ell = hyp_param['ell'] if (('ell' in hyp_param) and (not hyp_param['ell'] is None)) else 0 hyp_ell1 = hyp_param['ell1'] if (('ell1' in hyp_param) and (not hyp_param['ell1'] is None)) else np.nan hyp_ell2 = hyp_param['ell2'] if (('ell2' in hyp_param) and (not hyp_param['ell2'] is None)) else np.nan hyp_omega = hyp_param['omega'] if (('omega' in hyp_param) and (not hyp_param['omega'] is None)) else 0 hyp_omega1 = hyp_param['omega1'] if (('omega1' in hyp_param) and (not hyp_param['omega1'] is None)) else np.nan hyp_omega2 = hyp_param['omega2'] if (('omega2' in hyp_param) and (not hyp_param['omega2'] is None)) else np.nan hyp_pi = hyp_param['pi'] if (('pi' in hyp_param) and (not hyp_param['pi'] is None)) else 0 hyp_nu = hyp_param['nu'] if (('nu' in hyp_param) and (not hyp_param['nu'] is None)) else 0 #select kernel function if k_type == 1: #independent cells #cell ids in training data # cid_train = df_nerg_cellatten.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values cid_train = df_nerg_cellatten.index.values #cell ids in prediction data # cid_prdct = df_cell_info.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values cid_prdct = df_cell_info.index.values #compute mean and cov of cell anelastic coeffs for prediction scenarios ca_prdct_mu, _, ca_prdct_cov = pylib_kern.PredictGroupKern(cid_prdct, cid_train, c_train_mu=ca_mean_train, c_train_sig=ca_sig_train, hyp_mean_c=hyp_mean_ca , hyp_omega=hyp_omega) if k_type == 2: #exponetial #cell coordinates of training data ct_train = df_nerg_cellatten.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values #cell coordinates of prediction data ct_prdct = df_cell_info.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values #compute mean and cov of cell anelastic coeffs for prediction scenarios ca_prdct_mu, _, ca_prdct_cov = pylib_kern.PredictExpKern(ct_prdct, ct_train, c_train_mu=ca_mean_train, c_train_sig=ca_sig_train, hyp_mean_c=hyp_mean_ca, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi) if k_type == 3: #squared exponetial #cell coordinates of training data ct_train = df_nerg_cellatten.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values #cell coordinates of prediction data ct_prdct = df_cell_info.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values #compute mean and cov of cell anelastic coeffs for prediction scenarios ca_prdct_mu, _, ca_prdct_cov = pylib_kern.PredictSqExpKern(ct_prdct, ct_train, c_train_mu=ca_mean_train, c_train_sig=ca_sig_train, hyp_mean_c=hyp_mean_ca, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi) if k_type == 4: #Matern #cell coordinates of training data ct_train = df_nerg_cellatten.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values #cell coordinates of prediction data ct_prdct = df_cell_info.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values #compute mean and cov of cell anelastic coeffs for prediction scenarios ca_prdct_mu, _, ca_prdct_cov = pylib_kern.PredictSqMaternKern(ct_prdct, ct_train, c_train_mu=ca_mean_train, c_train_sig=ca_sig_train, hyp_mean_c=hyp_mean_ca, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi) if k_type == 5: #exponetial and spatially independent composite #cell coordinates of training data ct_train = df_nerg_cellatten.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values #cell coordinates of prediction data ct_prdct = df_cell_info.loc[:,nerg_catten_info[nerg_ca]['cor_info']].values #compute mean and cov of cell anelastic coeffs for prediction scenarios ca_prdct_mu, _, ca_prdct_cov = pylib_kern.PredictNegExpSptInptKern(ct_prdct, ct_train, c_train_mu=ca_mean_train, c_train_sig=ca_sig_train, hyp_mean_c=hyp_mean_ca, hyp_ell1=hyp_ell1, hyp_omega1=hyp_omega1, hyp_omega2=hyp_omega2, hyp_pi=hyp_pi) #sample cell-specific anelastic coefficients for prediction scenarios ca_prdct_samp = MVNRnd(mean=ca_prdct_mu, cov=ca_prdct_cov, n_samp=n_samp) ca_prdct_sig = np.sqrt( np.diag(ca_prdct_cov) ) #effect of anelastic attenuation nerg_atten_prdct_samp += cell_path @ ca_prdct_samp nerg_atten_prdct_mu += cell_path @ ca_prdct_mu nerg_atten_prdct_var += np.square(cell_path) @ ca_prdct_sig**2 #summarize individual anelastic components nerg_atten_cmp[nerg_ca] = [cell_path @ ca_prdct_mu, np.sqrt(np.square(cell_path) @ ca_prdct_sig**2), cell_path @ ca_prdct_samp] #total non-ergodic effects nerg_effects_prdct_samp = nerg_vcm_prdct_samp + nerg_atten_prdct_samp nerg_effects_prdct_mu = nerg_vcm_prdct_mu + nerg_atten_prdct_mu nerg_effects_prdct_sig = np.sqrt(nerg_vcm_prdct_var + nerg_atten_prdct_var) return nerg_effects_prdct_samp, nerg_vcm_prdct_samp, nerg_atten_prdct_samp, \ nerg_effects_prdct_mu, nerg_effects_prdct_sig, \ nerg_vcm_cmp, nerg_atten_cmp # Multivariate normal distribution random samples #-------------------------------------- def MVNRnd(mean = None, cov = None, seed = None, n_samp = None, flag_sp = False, flag_list = False): ''' Draw random samples from a Multivariable Normal distribution Parameters ---------- mean : np.array(n), optional Mean array. The default is None. cov : np.array(n,n), optional Covariance Matrix. The default is None. seed : int, optional Seed number of random number generator. The default is None. n_samp : int, optional Number of samples. The default is None. flag_sp : boolean, optional Sparse covariance matrix flag; if sparse flag_sp = True. The default is False. flag_list : boolean, optional Flag returning output as list. The default is False. Returns ------- samp Sampled values. ''' #if not already covert to list if flag_list: seed_list = seed if not seed is None else [None] else: seed_list = [seed] #number of dimensions n_dim = len(mean) if not mean is None else cov.shape[0] assert(cov.shape == (n_dim,n_dim)),'Error. Inconsistent size of mean array and covariance matrix' #set mean array to zero if not given if mean is None: mean = np.zeros(n_dim) #compute L D L' decomposition if flag_sp: cov = cov.toarray() L, D, _ = scipylinalg.ldl(cov) assert( not np.count_nonzero(D - np.diag(np.diagonal(D))) ),'Error. D not diagonal' assert( np.all(np.diag(D) > -1e-1) ),'Error. D diagonal is negative' #extract diagonal from D matrix, set to zero any negative entries due to bad conditioning d = np.diagonal(D).copy() d[d<0] = 0 #compute Q matrix Q = L @ np.diag(np.sqrt(d)) #generate random sample samp_list = list() for k, seed in enumerate(seed_list): #genereate seed numbers if not given if seed is None: seed = np.random.standard_normal(size=(n_dim, n_samp)) #generate random multi-normal random samples samp = Q @ (seed ) samp += mean[:,np.newaxis] if samp.ndim > 1 else mean #summarize samples samp_list.append( samp ) return samp_list if flag_list else samp_list[0]
21,359
54.051546
154
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/ground_motions/pylib_cell_dist.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun May 3 17:25:10 2020 @author: glavrent """ #load libraries import numpy as np import geopy.distance as geopydist def ComputeDistUnGridCells(pt1, pt2, cells, diffx, diffy, flagUTM=False): ''' Compute the path distances of uniformly gridded cells Parameters ---------- pt1 : np.array(3) Latitude, Longitude, elevation coordinates of first point. pt2 : np.array(3) Latitude, Longitude, elevation coordinates of second point. cells : np.array(n_cells, 4) Cell coordinates: Cartesian or LatLon Latitude, Longitude, bottom and top elevation of cells [x, y, elev_bot, elev_top] Lat Lon coordinates: Latitude, Longitude, bottom and top elevation of cells [lon, lat, elev_bot, elev_top] diffx : real Latitude interval of cells. diffy : real Longitude interval of cells. Returns ------- dm : np.array(n_cells) Distance path on each cell. ''' #import pdb; pdb.set_trace() #grid points x_grid = np.unique(cells[:, 0]) y_grid = np.unique(cells[:, 1]) z_grid = np.unique(cells[:, 2]) ## find x,y,z grid points which are between source and site x_v = np.sort([pt1[0], pt2[0]]) x_g_pts = x_grid[(x_v[0] <= x_grid) & (x_grid < x_v[1])] y_v = np.sort([pt1[1], pt2[1]]) y_g_pts = y_grid[(y_v[0] <= y_grid) & (y_grid < y_v[1])] z_v = np.sort([pt1[2], pt2[2]]) z_g_pts = z_grid[(z_v[0] <= z_grid) & (z_grid < z_v[1])] #p1-pt2 vector vec = np.subtract(pt1, pt2) # intersection points for x normal = [1, 0, 0]; ptx = np.ones(len(x_g_pts) * 3) if len(x_g_pts) > 0: ptx = ptx.reshape(len(x_g_pts), 3) for i, xv in enumerate(x_g_pts): ptplane = [xv, y_grid[0], 0] d = np.divide(np.dot(np.subtract(ptplane,pt1),normal), np.dot(vec,normal)) pt = pt1 + d * vec ptx[i] = pt else: ptx = [[-999, -999, -999]] # intersection points for y normal = [0, 1, 0]; pty = np.ones(len(y_g_pts) * 3) if len(y_g_pts) > 0: pty = pty.reshape(len(y_g_pts), 3) for i, yv in enumerate(y_g_pts): ptplane = [x_grid[0], yv, 0] d = np.divide(np.dot(np.subtract(ptplane,pt1),normal), np.dot(vec,normal)) pt = pt1 + d * vec pty[i] = pt else: pty = [[-999, -999, -999]] # intersection points for z normal = [0, 0, 1] ptz = np.ones(len(z_g_pts) * 3) if len(z_g_pts) > 0: ptz = ptz.reshape(len(z_g_pts), 3) for i, zv in enumerate(z_g_pts): ptplane = [x_grid[0], y_grid[0], zv] d = np.divide(np.dot(np.subtract(ptplane,pt1),normal), np.dot(vec,normal)) pt = pt1 + d * vec ptz[i] = pt else: ptz = [[-999, -999, -999]] #summarize all intersection points ptall = np.concatenate(([pt1], [pt2], ptx, pty, ptz)) ptall = ptall[(ptall[:, 0] != -999) & (ptall[:, 1] != -999) & (ptall[:, 2] != -999)] ptall = np.unique(ptall, axis=0) if pt1[0] != pt2[0]: ptall = ptall[ptall[:, 0].argsort()] #sort points by x coordinate else: ptall = ptall[ptall[:, 1].argsort()] #sort points by y coordinate #cell ids id_cells = np.arange(len(cells)) #compute cell distance idx = np.zeros(len(ptall)-1) distances = np.ones(len(ptall)-1) for i in range(len(ptall) - 1): p1 = ptall[i] #first intersection point p2 = ptall[i+1] #second intersection point #cell indices of cells where the first intersection point belongs idx1 = id_cells[(cells[:, 0] <= p1[0]) & (p1[0] <= cells[:, 0] + diffx) & \ (cells[:, 1] <= p1[1]) & (p1[1] <= cells[:, 1] + diffy) & \ (cells[:, 2] <= p1[2]) & (p1[2] <= cells[:, 3])] #cell indices of cells where the second intersection point belongs idx2 = id_cells[(cells[:, 0] <= p2[0]) & (p2[0] <= cells[:, 0] + diffx) & \ (cells[:, 1] <= p2[1]) & (p2[1] <= cells[:, 1] + diffy) & \ (cells[:, 2] <= p2[2]) & (p2[2] <= cells[:, 3])] #common indices of first and second int points idx[i] = np.intersect1d(idx1, idx2) #compute path distance if not flagUTM: dxy = geopydist.distance(ptall[i,(1,0)],ptall[i + 1,(1,0)]).km else: dxy = np.linalg.norm(ptall[i,0:1] - ptall[i + 1,0:1]) dz = ptall[i,2] - ptall[i + 1,2] distances[i] = np.sqrt(dxy** 2 + dz ** 2) dm = np.zeros(len(cells)) dm[idx.astype(int)] = distances return dm def ComputeDistGridCells(pt1, pt2, cells, flagUTM=False): ''' Compute the path distances of gridded cells Parameters ---------- pt1 : np.array(3) Latitude, Longitude, elevation coordinates of first point. pt2 : np.array(3) Latitude, Longitude, elevation coordinates of second point. cells : np.array(n_cells, 6) Latitude, Longitude, elevation of bottom left (q1) and top right (q8) corrners of cells [q1_lat, q1_lon, q1_elev, q8_lat, q8_lon, q8_elev] diffx : real Latitude interval of cells. diffy : real Longitude interval of cells. Returns ------- dm : np.array(n_cells) Distance path on each cell. ''' #import pdb; pdb.set_trace() #grid points x_grid = np.unique(cells[:, 0]) y_grid = np.unique(cells[:, 1]) z_grid = np.unique(cells[:, 2]) ## find x,y,z grid points which are between source and site x_v = np.sort([pt1[0], pt2[0]]) x_g_pts = x_grid[(x_v[0] <= x_grid) & (x_grid < x_v[1])] y_v = np.sort([pt1[1], pt2[1]]) y_g_pts = y_grid[(y_v[0] <= y_grid) & (y_grid < y_v[1])] z_v = np.sort([pt1[2], pt2[2]]) z_g_pts = z_grid[(z_v[0] <= z_grid) & (z_grid < z_v[1])] #p1-pt2 vector vec = np.subtract(pt1, pt2) # intersection points for x normal = [1, 0, 0]; ptx = np.ones(len(x_g_pts) * 3) if len(x_g_pts) > 0: ptx = ptx.reshape(len(x_g_pts), 3) for i, xv in enumerate(x_g_pts): ptplane = [xv, y_grid[0], 0] d = np.divide(np.dot(np.subtract(ptplane,pt1),normal), np.dot(vec,normal)) pt = pt1 + d * vec ptx[i] = pt else: ptx = [[-999, -999, -999]] # intersection points for y normal = [0, 1, 0]; pty = np.ones(len(y_g_pts) * 3) if len(y_g_pts) > 0: pty = pty.reshape(len(y_g_pts), 3) for i, yv in enumerate(y_g_pts): ptplane = [x_grid[0], yv, 0] d = np.divide(np.dot(np.subtract(ptplane,pt1),normal), np.dot(vec,normal)) pt = pt1 + d * vec pty[i] = pt else: pty = [[-999, -999, -999]] # intersection points for z normal = [0, 0, 1] ptz = np.ones(len(z_g_pts) * 3) if len(z_g_pts) > 0: ptz = ptz.reshape(len(z_g_pts), 3) for i, zv in enumerate(z_g_pts): ptplane = [x_grid[0], y_grid[0], zv] d = np.divide(np.dot(np.subtract(ptplane,pt1),normal), np.dot(vec,normal)) pt = pt1 + d * vec ptz[i] = pt else: ptz = [[-999, -999, -999]] #summarize all intersection points ptall = np.concatenate(([pt1], [pt2], ptx, pty, ptz)) ptall = ptall[(ptall[:, 0] != -999) & (ptall[:, 1] != -999) & (ptall[:, 2] != -999)] #ptall = np.unique(ptall.round, axis=0, return_index=True) _, i_ptall_unq = np.unique(ptall.round(decimals=7), axis=0, return_index=True) ptall = ptall[i_ptall_unq,:] # if pt1[0] != pt2[0]: if abs(pt1[0] - pt2[0]) > 1e-6: ptall = ptall[ptall[:, 0].argsort()] else: ptall = ptall[ptall[:, 1].argsort()] #compute cell distance id_cells = np.arange(len(cells)) idx = np.ones(len(ptall)-1) distances = np.ones(len(ptall)-1) for i in range(len(ptall) - 1): p1 = ptall[i] #first intersection point p2 = ptall[i+1] #second intersection point #cell indices where the first point belongs tol = 1e-9 idx1 = id_cells[(cells[:, 0]-tol <= p1[0]) & (p1[0] <= cells[:, 3]+tol) & \ (cells[:, 1]-tol <= p1[1]) & (p1[1] <= cells[:, 4]+tol) & \ (cells[:, 2]-tol <= p1[2]) & (p1[2] <= cells[:, 5]+tol)] #cell indices where the second point belongs idx2 = id_cells[(cells[:, 0]-tol <= p2[0]) & (p2[0] <= cells[:, 3]+tol) & \ (cells[:, 1]-tol <= p2[1]) & (p2[1] <= cells[:, 4]+tol) & \ (cells[:, 2]-tol <= p2[2]) & (p2[2] <= cells[:, 5]+tol)] #common indices of first and second point try: idx[i] = np.intersect1d(idx1, idx2) except ValueError: print('i_pt: ', i) print('idx1: ', idx1) print('idx2: ', idx2) print('p1: ', p1) print('p2: ', p2) # import pdb; pdb.set_trace() raise #compute path distance if not flagUTM: dxy = geopydist.distance(ptall[i,(1,0)],ptall[i + 1,(1,0)]).km else: dxy = np.linalg.norm(ptall[i,0:2] - ptall[i + 1,0:2]) dz = ptall[i,2] - ptall[i + 1,2] distances[i] = np.sqrt(dxy** 2 + dz ** 2) dm = np.zeros(len(cells)) dm[idx.astype(int)] = distances return dm
9,551
32.75265
95
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/ground_motions/pylib_kernels.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sat Aug 20 13:52:51 2022 @author: glavrent """ # Packages # --------------------------- #arithmetic libraries import numpy as np from scipy import linalg as scipylinalg from sklearn.gaussian_process.kernels import Matern # Kernel Functions #-------------------------------------- # group kernel function #--- --- --- --- --- --- --- --- def KernelGroup(grp_1, grp_2, hyp_omega = 0, delta = 1e-9): ''' Compute kernel function for perfect correlation between group variables Parameters ---------- grp_1 : np.array IDs for first group. grp_2 : np.array IDs for second group. hyp_omega : non-negative real, optional Scale of kernel function. The default is 0. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- cov_mat : np.array Covariance Matrix. ''' #tolerance for station id comparison r_tol = np.min([0.01/np.max([np.abs(grp_1).max(), np.abs(grp_2).max()]), 1e-11]) #number of grid nodes n_pt_1 = grp_1.shape[0] n_pt_2 = grp_2.shape[0] #number of dimensions n_dim = grp_1.ndim #create cov. matrix cov_mat = np.zeros([n_pt_1,n_pt_2]) #initialize if n_dim == 1: for i in range(n_pt_1): cov_mat[i,:] = hyp_omega**2 * np.isclose(grp_1[i], grp_2, rtol=r_tol).flatten() else: for i in range(n_pt_1): cov_mat[i,:] = hyp_omega**2 * (scipylinalg.norm(grp_1[i] - grp_2, axis=1) < r_tol) if n_pt_1 == n_pt_2: for i in range(n_pt_1): cov_mat[i,i] += delta return cov_mat # exponential kernel #--- --- --- --- --- --- --- --- def KernelExp(t_1, t_2, hyp_ell = 0, hyp_omega = 0, hyp_pi = 0, delta = 1e-9): ''' Compute exponential kernel function Parameters ---------- t_1 : np.array Coordinates of first group. t_2 : np.array Coordinates of second group. hyp_ell : non-negative real, optional Correlation length. The default is 0. hyp_omega : non-negative real, optional Scale of kernel function. The default is 0. hyp_pi : non-negative real, optional Constant of kernel function. The default is 0. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- cov_mat : np.array Covariance Matrix. ''' #number of grid nodes n_pt_1 = t_1.shape[0] n_pt_2 = t_2.shape[0] #number of dimensions n_dim = t_1.ndim #create cov. matrix cov_mat = np.zeros([n_pt_1,n_pt_2]) #initialize for i in range(n_pt_1): dist = scipylinalg.norm(t_1[i] - t_2,axis=1) if n_dim > 1 else np.abs(t_1[i] - t_2) cov_mat[i,:] = hyp_pi**2 + hyp_omega**2 * np.exp(- dist/hyp_ell) if n_pt_1 == n_pt_2: for i in range(n_pt_1): cov_mat[i,i] += delta return cov_mat # squared exponential kernel #--- --- --- --- --- --- --- --- def KernelSqExp(t_1, t_2, hyp_ell = 0, hyp_omega = 0, hyp_pi = 0, delta = 1e-9): ''' Compute squared exponential kernel function Parameters ---------- t_1 : np.array Coordinates of first group. t_2 : np.array Coordinates of second group. hyp_ell : non-negative real, optional Correlation length. The default is 0. hyp_omega : non-negative real, optional Scale of kernel function. The default is 0. hyp_pi : non-negative real, optional Constant of kernel function. The default is 0. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- cov_mat : np.array Covariance Matrix. ''' #number of grid nodes n_pt_1 = t_1.shape[0] n_pt_2 = t_2.shape[0] #number of dimensions n_dim = t_1.ndim #create cov. matrix cov_mat = np.zeros([n_pt_1,n_pt_2]) #initialize for i in range(n_pt_1): dist = scipylinalg.norm(t_1[i] - t_2,axis=1) if n_dim > 1 else np.abs(t_1[i] - t_2) cov_mat[i,:] = hyp_pi**2 + hyp_omega**2 * np.exp(- dist**2/hyp_ell**2) if n_pt_1 == n_pt_2: for i in range(n_pt_1): cov_mat[i,i] += delta return cov_mat # matern exponential kernel #--- --- --- --- --- --- --- --- def MaternKernel(t_1, t_2, hyp_ell = 0, hyp_omega = 0, hyp_pi = 0, hyp_nu=1.5, delta = 1e-9): ''' Compute Matern kernel function Parameters ---------- t_1 : np.array Coordinates of first group. t_2 : np.array Coordinates of second group. hyp_ell : non-negative real, optional Correlation length. The default is 0. hyp_omega : non-negative real, optional Scale of kernel function. The default is 0. hyp_pi : non-negative real, optional Constant of kernel function. The default is 0. hyp_nu : non-negative real, optional Smoothness parameter. The default is 1.5. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- cov_mat : np.array Covariance Matrix. ''' #number of grid nodes n_pt_1 = t_1.shape[0] n_pt_2 = t_2.shape[0] #number of dimensions n_dim = t_1.ndim #distance matrix dist_mat = np.array([scipylinalg.norm(t1 - t_2, axis=1) if n_dim > 1 else np.abs(t1 - t_2) for t1 in t_1]) #create cov. matrix cov_mat = hyp_omega**2 * Matern(nu=hyp_nu, length_scale=hyp_ell)(0, dist_mat.ravel()[:, np.newaxis]).reshape(dist_mat.shape) cov_mat += hyp_pi**2 if n_pt_1 == n_pt_2: for i in range(n_pt_1): cov_mat[i,i] += delta return cov_mat # composite exponential kernel and spatially independent #--- --- --- --- --- --- --- --- def KernelNegExpSptInpt(t_1, t_2, hyp_ell1 = 0, hyp_omega1 = 0, hyp_omega2 = 0, hyp_pi = 0, delta = 1e-9): ''' Compute composite kernel function, with negative exponential and spatially idependent components Parameters ---------- t_1 : np.array Coordinates of first group. t_2 : np.array Coordinates of second group. hyp_ell1 : non-negative real, optional Correlation length of neg. exponential component. The default is 0. hyp_omega1 : non-negative real, optional Scale of neg. exponential component. The default is 0. hyp_omega2 : non-negative real, optional Scale of spatially independent component. The default is 0. hyp_pi : non-negative real, optional Constant of kernel function. The default is 0. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- cov_mat : TYPE DESCRIPTION. ''' #number of grid nodes n_pt_1 = t_1.shape[0] n_pt_2 = t_2.shape[0] #negative exponetial component cov_mat = KernelExp(t_1, t_2, hyp_ell=hyp_ell1, hyp_omega=hyp_omega1, hyp_pi=hyp_pi, delta=1e-9) #spatially independent component cov_mat += KernelGroup(t_1, t_2, hyp_omega=hyp_omega2, delta=0) return cov_mat # Predictive Functions #-------------------------------------- # predict coeffs with group kernel function #--- --- --- --- --- --- --- --- def PredictGroupKern(g_prdct, g_train, c_train_mu, c_train_sig = None, hyp_mean_c = 0, hyp_omega = 0, delta = 1e-9): ''' Predict conditional coefficients based on group kernel function. Parameters ---------- g_prdct : np.array Group IDs of prediction cases. g_train : np.array Group IDs of training cases. c_train_mu : np.array Mean values of non-ergodic coefficient of training cases. c_train_sig : np.array, optional Standard deviations of non-ergodic coefficient of training cases. The default is None. hyp_mean_c : real, optional Mean of non-ergodic coefficient. The default is 0. hyp_omega : non-negative real, optional Scale of kernel function. The default is 0. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- c_prdct_mu : np.array Mean value of non-ergodic coefficient for prediction cases. c_prdct_sig : np.array Standard deviations of non-ergodic coefficient for prediction cases. c_prdct_cov : np.array Covariance matrix of non-ergodic coefficient for prediction cases. ''' #remove mean effect from training coefficients c_train_mu = c_train_mu - hyp_mean_c #uncertainty in training data if c_train_sig is None: c_train_sig = np.zeros(len(c_train_mu)) c_train_cov = np.diag(c_train_sig**2) if c_train_sig.ndim == 1 else c_train_sig #covariance between training data K = KernelGroup(g_train, g_train, hyp_omega=hyp_omega, delta=delta) #covariance between data and new locations k = KernelGroup(g_prdct, g_train, hyp_omega=hyp_omega, delta=0) #covariance between new locations k_star = KernelGroup(g_prdct, g_prdct, hyp_omega=hyp_omega, delta=0) #inverse of covariance matrix K_inv = scipylinalg.inv(K) #product of k * K^-1 kK_inv = k.dot(K_inv) #posterior mean and variance at new locations c_prdct_mu = kK_inv.dot(c_train_mu) c_prdct_cov = k_star - kK_inv.dot(k.transpose()) + kK_inv.dot( c_train_cov.dot(kK_inv.transpose()) ) #posterior standard dev. at new locations c_prdct_sig = np.sqrt(np.diag(c_prdct_cov)) #add mean effect from training coefficients c_prdct_mu += hyp_mean_c return c_prdct_mu, c_prdct_sig, c_prdct_cov # predict coeffs with exponential kernel #--- --- --- --- --- --- --- --- def PredictExpKern(t_prdct, t_train, c_train_mu, c_train_sig = None, hyp_mean_c = 0, hyp_ell = 0, hyp_omega = 0, hyp_pi = 0, delta = 1e-9): ''' Predict conditional coefficients based on exponential kernel function. Parameters ---------- t_prdct : np.array Coordinates of prediction cases. t_train : np.array Coordinates of training cases. c_train_mu : np.array Mean values of non-ergodic coefficient of training cases. c_train_sig : np.array, optional Standard deviations of non-ergodic coefficient of training cases. The default is None. hyp_mean_c : real, optional Mean of non-ergodic coefficient. The default is 0. hyp_ell : non-negative real, optional Correlation length of kernel function.. The default is 0. hyp_omega : non-negative real, optional Scale of kernel function. The default is 0. hyp_pi : postive real, optional Constant of kernel function. The default is 0. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- c_prdct_mu : np.array Mean value of non-ergodic coefficient for prediction cases. c_prdct_sig : np.array Standard deviations of non-ergodic coefficient for prediction cases. c_prdct_cov : np.array Covariance matrix of non-ergodic coefficient for prediction cases. ''' #remove mean effect from training coefficients c_train_mu = c_train_mu - hyp_mean_c #uncertainty in training data if c_train_sig is None: c_train_sig = np.zeros(len(c_train_mu)) c_train_cov = np.diag(c_train_sig**2) if c_train_sig.ndim == 1 else c_train_sig #covariance between training data K = KernelExp(t_train, t_train, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, delta=delta) #covariance between data and new locations k = KernelExp(t_prdct, t_train, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, delta=0) #covariance between new locations k_star = KernelExp(t_prdct, t_prdct, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, delta=0) #inverse of covariance matrix K_inv = scipylinalg.inv(K) #product of k * K^-1 kK_inv = k.dot(K_inv) #posterior mean and variance at new locations c_prdct_mu = kK_inv.dot(c_train_mu) c_prdct_cov = k_star - kK_inv.dot(k.transpose()) + kK_inv.dot( c_train_cov.dot(kK_inv.transpose()) ) #posterior standard dev. at new locations c_prdct_sig = np.sqrt(np.diag(c_prdct_cov)) #add mean effect from training coefficients c_prdct_mu += hyp_mean_c return c_prdct_mu, c_prdct_sig, c_prdct_cov # predict coeffs with squared exponential kernel #--- --- --- --- --- --- --- --- def PredictSqExpKern(t_prdct, t_train, c_train_mu, c_train_sig = None, hyp_mean_c = 0, hyp_ell = 0, hyp_omega = 0, hyp_pi = 0, delta = 1e-9): ''' Predict conditional coefficients based on squared exponential kernel function. Parameters ---------- t_prdct : np.array Coordinates of prediction cases. t_train : np.array Coordinates of training cases. c_train_mu : np.array Mean values of non-ergodic coefficient of training cases. c_train_sig : np.array, optional Standard deviations of non-ergodic coefficient of training cases. The default is None. hyp_mean_c : real, optional Mean of non-ergodic coefficient. The default is 0. hyp_ell : non-negative real, optional Correlation length of kernel function.. The default is 0. hyp_omega : non-negative real, optional Scale of kernel function. The default is 0. hyp_pi : postive real, optional Constant of kernel function. The default is 0. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- c_prdct_mu : np.array Mean value of non-ergodic coefficient for prediction cases. c_prdct_sig : np.array Standard deviations of non-ergodic coefficient for prediction cases. c_prdct_cov : np.array Covariance matrix of non-ergodic coefficient for prediction cases. ''' #remove mean effect from training coefficients c_train_mu = c_train_mu - hyp_mean_c #uncertainty in training data if c_train_sig is None: c_train_sig = np.zeros(len(c_train_mu)) c_train_cov = np.diag(c_train_sig**2) if c_train_sig.ndim == 1 else c_train_sig #covariance between training data K = KernelNegExpSptInpt(t_train, t_train, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, delta=delta) #covariance between data and new locations k = KernelNegExpSptInpt(t_prdct, t_train, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, delta=0) #covariance between new locations k_star = KernelNegExpSptInpt(t_prdct, t_prdct, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, delta=0) #inverse of covariance matrix K_inv = scipylinalg.inv(K) #product of k * K^-1 kK_inv = k.dot(K_inv) #posterior mean and variance at new locations c_prdct_mu = kK_inv.dot(c_train_mu) c_prdct_cov = k_star - kK_inv.dot(k.transpose()) + kK_inv.dot( c_train_cov.dot(kK_inv.transpose()) ) #posterior standard dev. at new locations c_prdct_sig = np.sqrt(np.diag(c_prdct_cov)) #add mean effect from training coefficients c_prdct_mu += hyp_mean_c return c_prdct_mu, c_prdct_sig, c_prdct_cov # predict coeffs with Matern kernel #--- --- --- --- --- --- --- --- def PredictMaternKern(t_prdct, t_train, c_train_mu, c_train_sig = None, hyp_mean_c = 0, hyp_ell = 0, hyp_omega = 0, hyp_pi = 0, hyp_nu=1.5, delta = 1e-9): ''' Predict conditional coefficients based on Matern kernel function. Parameters ---------- t_prdct : np.array Coordinates of prediction cases. t_train : np.array Coordinates of training cases. c_train_mu : np.array Mean values of non-ergodic coefficient of training cases. c_train_sig : np.array, optional Standard deviations of non-ergodic coefficient of training cases. The default is None. hyp_mean_c : real, optional Mean of non-ergodic coefficient. The default is 0. hyp_ell : non-negative real, optional Correlation length of kernel function.. The default is 0. hyp_omega : non-negative real, optional Scale of kernel function. The default is 0. hyp_pi : postive real, optional Constant of kernel function. The default is 0. hyp_nu: positive real, optional Smoothness parameter. The default is 1.5. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- c_prdct_mu : np.array Mean value of non-ergodic coefficient for prediction cases. c_prdct_sig : np.array Standard deviations of non-ergodic coefficient for prediction cases. c_prdct_cov : np.array Covariance matrix of non-ergodic coefficient for prediction cases. ''' #remove mean effect from training coefficients c_train_mu = c_train_mu - hyp_mean_c #uncertainty in training data if c_train_sig is None: c_train_sig = np.zeros(len(c_train_mu)) c_train_cov = np.diag(c_train_sig**2) if c_train_sig.ndim == 1 else c_train_sig #covariance between training data K = MaternKernel(t_train, t_train, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, hyp_nu=hyp_nu, delta=delta) #covariance between data and new locations k = MaternKernel(t_prdct, t_train, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, hyp_nu=hyp_nu, delta=0) #covariance between new locations k_star = MaternKernel(t_prdct, t_prdct, hyp_ell=hyp_ell, hyp_omega=hyp_omega, hyp_pi=hyp_pi, hyp_nu=hyp_nu, delta=0) #inverse of covariance matrix K_inv = scipylinalg.inv(K) #product of k * K^-1 kK_inv = k.dot(K_inv) #posterior mean and variance at new locations c_prdct_mu = kK_inv.dot(c_train_mu) c_prdct_cov = k_star - kK_inv.dot(k.transpose()) + kK_inv.dot( c_train_cov.dot(kK_inv.transpose()) ) #posterior standard dev. at new locations c_prdct_sig = np.sqrt(np.diag(c_prdct_cov)) #add mean effect from training coefficients c_prdct_mu += hyp_mean_c return c_prdct_mu, c_prdct_sig, c_prdct_cov # predict coeffs with composite exponential and spatially independent kernel function #--- --- --- --- --- --- --- --- def PredictNegExpSptInptKern(t_prdct, t_train, c_train_mu, c_train_sig = None, hyp_mean_c = 0, hyp_ell1 = 0, hyp_omega1 = 0, hyp_omega2 = 0, hyp_pi = 0, delta = 1e-9): ''' Predict conditional coefficients based on composite exponential and spatially independent kernel function. Parameters ---------- t_prdct : np.array Coordinates of prediction cases. t_train : np.array Coordinates of training cases. c_train_mu : np.array Mean values of non-ergodic coefficient of training cases. c_train_sig : np.array, optional Standard deviations of non-ergodic coefficient of training cases. The default is None. hyp_mean_c : real, optional Mean of non-ergodic coefficient. The default is 0. hyp_ell1 : non-negative real, optional Correlation length of negative exponential kernel function. The default is 0. hyp_omega1 : non-negative real, optional Scale of negative exponential kernel function. The default is 0. hyp_omega2 : non-negative real, optional Scale of spatially independent kernel function. The default is 0. hyp_pi : postive real, optional Constant of kernel function. The default is 0. delta : non-negative real, optional Diagonal widening. The default is 1e-9. Returns ------- c_prdct_mu : np.array Mean value of non-ergodic coefficient for prediction cases. c_prdct_sig : np.array Standard deviations of non-ergodic coefficient for prediction cases. c_prdct_cov : np.array Covariance matrix of non-ergodic coefficient for prediction cases. ''' #remove mean effect from training coefficients c_train_mu = c_train_mu - hyp_mean_c #uncertainty in training data if c_train_sig is None: c_train_sig = np.zeros(len(c_train_mu)) c_train_cov = np.diag(c_train_sig**2) if c_train_sig.ndim == 1 else c_train_sig #covariance between training data K = KernelNegExpSptInpt(t_train, t_train, hyp_ell1=hyp_ell1, hyp_omega1=hyp_omega1, hyp_omega2=hyp_omega2, hyp_pi=hyp_pi, delta=delta) #covariance between data and new locations k = KernelNegExpSptInpt(t_prdct, t_train, hyp_ell1=hyp_ell1, hyp_omega1=hyp_omega1, hyp_omega2=hyp_omega2, hyp_pi=hyp_pi, delta=0) #covariance between new locations k_star = KernelNegExpSptInpt(t_prdct, t_prdct, hyp_ell1=hyp_ell1, hyp_omega1=hyp_omega1, hyp_omega2=hyp_omega2, hyp_pi=hyp_pi, delta=0) #inverse of covariance matrix K_inv = scipylinalg.inv(K) #product of k * K^-1 kK_inv = k.dot(K_inv) #posterior mean and variance at new locations c_prdct_mu = kK_inv.dot(c_train_mu) c_prdct_cov = k_star - kK_inv.dot(k.transpose()) + kK_inv.dot( c_train_cov.dot(kK_inv.transpose()) ) #posterior standard dev. at new locations c_prdct_sig = np.sqrt(np.diag(c_prdct_cov)) #add mean effect from training coefficients c_prdct_mu += hyp_mean_c return c_prdct_mu, c_prdct_sig, c_prdct_cov
21,651
34.966777
128
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/QGIS/pylib_QGIS.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue May 19 11:04:00 2020 @author: glavrent """ #load libraries #load GIS from qgis.core import QgsVectorLayer, QgsPointXY from qgis.core import QgsField, QgsFeature, QgsGeometry, QgsVectorFileWriter, QgsFeatureSink from qgis.PyQt.QtCore import QVariant def EQLayer(eq_data): ''' Create earthquake source layer for QGIS Parameters ---------- eq_data : pd.dataframe Dataframe for rupture points with fields: eqid, region, mag, SOF, Ztor, eqLat, eqLon Returns ------- eq_layer : TYPE QGIS layer with earthquake sources. ''' #create qgis layer for earthquake sources eq_layer = QgsVectorLayer("Point", "eq_pts", "memory") eq_pr = eq_layer.dataProvider() eq_pr.addAttributes([QgsField("eqid", QVariant.Int), QgsField("region", QVariant.Int), QgsField("mag", QVariant.Double), QgsField("SOF", QVariant.Int), QgsField("Ztor", QVariant.Double), QgsField("eqLat", QVariant.Double), QgsField("eqLon", QVariant.Double)]) #iterate over earthquakes, add on layer eq_layer.startEditing() for eq in eq_data.iterrows(): #earthquake info eq_info = eq[1][['eqid','region','mag','SOF','Ztor']].tolist() eq_latlon = eq[1][['eqLat','eqLon']].tolist() #define feature, earthquake eq_f = QgsFeature() eq_f.setGeometry(QgsGeometry.fromPointXY(QgsPointXY(eq_latlon[1],eq_latlon[0]))) eq_f.setAttributes(eq_info + eq_latlon) #add earthquake in layer eq_pr.addFeatures([eq_f]) #commit changes eq_layer.commitChanges() #update displacement layer eq_layer.updateExtents() return eq_layer def STALayer(sta_data): ''' Create station layer for QGIS Parameters ---------- sta_data : pd.dataframe Dataframe for rupture points with fields: 'ssn','region','Vs30','Z1.0','StaLat','StaLon' eqid','region','mag','SOF','eqLat','eqLon' Returns ------- sta_layer : TYPE QGIS layer with station points. ''' #create qgis layer for station locations sta_layer = QgsVectorLayer("Point", "sta_pts", "memory") sta_pr = sta_layer.dataProvider() sta_pr.addAttributes([QgsField("ssn", QVariant.Int), QgsField("region", QVariant.Int), QgsField("Vs30", QVariant.Double), QgsField("Z1.0", QVariant.Double), QgsField("staLat", QVariant.Double), QgsField("staLon", QVariant.Double)]) #iterate over station, add on layer sta_layer.startEditing() for sta in sta_data.iterrows(): #earthquake info sta_info = sta[1][['ssn','region','Vs30','Z1.0']].tolist() sta_latlon = sta[1][['staLat','staLon']].tolist() #define feature, earthquake sta_f = QgsFeature() sta_f.setGeometry(QgsGeometry.fromPointXY(QgsPointXY(sta_latlon[1],sta_latlon[0]))) sta_f.setAttributes(sta_info + sta_latlon) #add earthquake in layer sta_pr.addFeatures([sta_f]) #commit changes sta_layer.commitChanges() #update displacement layer sta_layer.updateExtents() return sta_layer
3,513
31.841121
92
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/QGIS/__init__.py
0
0
0
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/pylib_stats.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 15 13:56:13 2022 @author: glavrent Other python statistics functions """ #imprort libraries import numpy as np def CalcRMS(samp_q, samp_p): ''' Compute root mean square error between observation samples (samp_p) and model samples (samp_p) Parameters ---------- samp_q : np.array() Model Samples. samp_p : np.array() Data Samples. Returns ------- real root mean square error ''' #errors e = samp_q - samp_p return np.sqrt(np.mean(e**2)) def CalcLKDivergece(samp_q, samp_p): ''' Compute Kullback–Leibler divergence of observation samples (samp_p) based on model samples (samp_p) Parameters ---------- samp_q : np.array() Model Samples. samp_p : np.array() Data Samples. Returns ------- real Kullback–Leibler divergence. ''' #create histogram bins _, hist_bins = np.histogram(np.concatenate([samp_p,samp_q])) #count of p and q distribution p, _ = np.histogram(samp_p, bins=hist_bins) q, _ = np.histogram(samp_q, bins=hist_bins) #remove bins empty in any dist, otherwise kl= +/- inf i_empty_bins = np.logical_or(p==0, q==0) p = p[~i_empty_bins] q = q[~i_empty_bins] #normalize to compute probabilites p = p/p.sum() q = q/q.sum() return sum(p[i] * np.log2(p[i]/q[i]) for i in range(len(p)))
1,508
19.671233
78
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/__init__.py
0
0
0
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/pystan/regression_pystan_model3_uncorr_cells_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 13 18:22:15 2021 @author: glavrent """ #load variables import os import pathlib import glob import re #regular expression package import pickle from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_2_erg=0, c_3_erg=0, c_a_erg=0, runstan_flag=True, n_iter=600, n_chains=4, adapt_delta=0.8, max_treedepth=10, pystan_ver=2, pystan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and uncorrelated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_2_erg : double, optional Value of ergodic geometrical spreading coefficient. The default is 0. c_3_erg : double, optional Value of ergodic Vs30 coefficient. The default is 0. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter : integer, optional Number of stan samples. The default is 600. n_chains : integer, optional Number of MCMC chains. The default is 4. runstan_flag : bool, optional Flag for running stan. If true run regression, if false read past regression output and summarize non-ergodic parameters. The default is True. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' #number of cores n_cpu = max(cpu_count() -1,1) ## Read Data #============================ #read stan model with open(stan_model_fname, "r") as f: stan_model_code = f.read() ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','x_3','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[3,4]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #geometrical spreading covariates x_2 = df_flatfile['x_2'].values #vs30 covariates x_3 = df_flatfile['x_3'].values[sta_idx] #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'rec_mu': np.zeros(y_data.shape), 'Y': y_data, 'x_2': x_2, 'x_3': x_3, 'c_2_erg': c_2_erg, 'c_3_erg': c_3_erg, 'c_a_erg': c_a_erg, 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'RC': celldist_valid.to_numpy(), } stan_data_fname = out_fname + '_stan_data' + '.Rdata' ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #filename for STAN regression raw output file saved as pkl stan_fit_fname = out_dir + out_fname + '_stan_fit' + '.pkl' #run stan if runstan_flag: #control paramters control_stan = {'adapt_delta':adapt_delta, 'max_treedepth':max_treedepth} if pystan_ver == 2: import pystan if (not pystan_parallel) or n_cpu<=n_chains: #compile stan_model = pystan.StanModel(model_code=stan_model_code) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=10, control = control_stan) else: #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #multi-processing arguments os.environ['STAN_NUM_THREADS'] = str(n_cpu_chain) extra_compile_args = ['-pthread', '-DSTAN_THREADS'] #compile stan_model = pystan.StanModel(model_code=stan_model_code, extra_compile_args=extra_compile_args) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=1, control = control_stan) elif pystan_ver == 3: import nest_asyncio import stan nest_asyncio.apply() #compile stan_model = stan.build(stan_model_code, data=stan_data, random_seed=1) #full Bayesian statistics stan_fit = stan_model.sample(num_chains=n_chains, num_samples=n_iter, max_depth=max_treedepth, delta=adapt_delta) #save stan model and fit pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) with open(stan_fit_fname, "wb") as f: pickle.dump({'model' : stan_model, 'fit' : stan_fit}, f, protocol=-1) else: #load model and fit for postprocessing if has already been executed with open(stan_fit_fname, "rb") as f: data_dict = pickle.load(f) stan_fit = data_dict['fit'] stan_model = data_dict['model'] del data_dict ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','mu_2p','mu_3s', 'ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'ell_2p', 'ell_3s', 'omega_2p', 'omega_3s', 'mu_cap', 'omega_cap', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_c_2p = ['c_2p.%i'%(k) for k in range(n_eq)] col_names_c_3s = ['c_3s.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = (col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_c_2p + col_names_c_3s + col_names_cap + col_names_dB) #summarize raw posterior distributions stan_posterior = np.stack([stan_fit[c_n].flatten() for c_n in col_names_hyp], axis=1) #adjustment terms if pystan_ver == 2: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_2p']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_3s']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB']), axis=1) else: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_2p'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_3s'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB'].T), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_c_2p, col_names_c_3s, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid.values @ cells_ca_mu cells_LcA_med = celldist_valid.values @ cells_ca_med cells_LcA_sig = np.sqrt(np.square(celldist_valid.values) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] #spatially varying geometrical spreading coefficient coeff_2p_mu = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].mean() for k in range(n_eq)]) coeff_2p_mu = coeff_2p_mu[eq_inv] coeff_2p_med = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].median() for k in range(n_eq)]) coeff_2p_med = coeff_2p_med[eq_inv] coeff_2p_sig = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].std() for k in range(n_eq)]) coeff_2p_sig = coeff_2p_sig[eq_inv] #spatially varying Vs30 coefficient coeff_3s_mu = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].mean() for k in range(n_sta)]) coeff_3s_mu = coeff_3s_mu[sta_inv] coeff_3s_med = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].median() for k in range(n_sta)]) coeff_3s_med = coeff_3s_med[sta_inv] coeff_3s_sig = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].std() for k in range(n_sta)]) coeff_3s_sig = coeff_3s_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #dataframe with flatfile info df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, coeff_2p_mu, coeff_3s_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, coeff_2p_med, coeff_3s_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, coeff_2p_sig, coeff_3s_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','c_2p_mean','c_3s_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'c_2p_med', 'c_3s_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'c_2p_sig', 'c_3s_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + coeff_2p_mu*x_2 + coeff_3s_mu*x_3[sta_inv] + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots for c_name in col_names_hyp: fig = stan_fit.traceplot(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '.png') #create trace plot with arviz ax = az.plot_trace(stan_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
21,718
46.944812
130
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/pystan/regression_pystan_model2_uncorr_cells_sparse_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 13 18:22:15 2021 @author: glavrent """ #load variables import os import pathlib import glob import re #regular expression package import pickle from joblib import cpu_count #arithmetic libraries import numpy as np from scipy import sparse #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_a_erg=0, runstan_flag=True, n_iter=600, n_chains=4, adapt_delta=0.8, max_treedepth=10, pystan_ver=2, pystan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and uncorrelated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter : integer, optional Number of stan samples. The default is 600. n_chains : integer, optional Number of MCMC chains. The default is 4. runstan_flag : bool, optional Flag for running stan. If true run regression, if false read past regression output and summarize non-ergodic parameters. The default is True. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Read Data #============================ #read stan model with open(stan_model_fname, "r") as f: stan_model_code = f.read() ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid].to_numpy() #cell-distance with only non-zero cells celldist_valid_sp = sparse.csr_matrix(celldist_valid) #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', np.abs(df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'NCELL_SP': len(celldist_valid_sp.data), 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'rec_mu': np.zeros(y_data.shape), 'RC_val': celldist_valid_sp.data, 'RC_w': celldist_valid_sp.indices+1, 'RC_u': celldist_valid_sp.indptr+1, 'c_a_erg': c_a_erg, 'Y': y_data, } stan_data_fname = out_fname + '_stan_data' + '.Rdata' ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #filename for STAN regression raw output file saved as pkl stan_fit_fname = out_dir + out_fname + '_stan_fit' + '.pkl' #run stan if runstan_flag: #control paramters control_stan = {'adapt_delta':adapt_delta, 'max_treedepth':max_treedepth} if pystan_ver == 2: import pystan if (not pystan_parallel) or n_cpu<=n_chains: #compile stan_model = pystan.StanModel(model_code=stan_model_code) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=10, control = control_stan) else: #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #multi-processing arguments os.environ['STAN_NUM_THREADS'] = str(n_cpu_chain) extra_compile_args = ['-pthread', '-DSTAN_THREADS'] #compile stan_model = pystan.StanModel(model_code=stan_model_code, extra_compile_args=extra_compile_args) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=1, control = control_stan) elif pystan_ver == 3: import nest_asyncio import stan nest_asyncio.apply() #compile stan_model = stan.build(stan_model_code, data=stan_data, random_seed=1) #full Bayesian statistics stan_fit = stan_model.sample(num_chains=n_chains, num_samples=n_iter, max_depth=max_treedepth, delta=adapt_delta) #save stan model and fit pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) with open(stan_fit_fname, "wb") as f: pickle.dump({'model' : stan_model, 'fit' : stan_fit}, f, protocol=-1) else: #load model and fit for postprocessing if has already been executed with open(stan_fit_fname, "rb") as f: data_dict = pickle.load(f) stan_fit = data_dict['fit'] stan_model = data_dict['model'] del data_dict ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'mu_cap', 'omega_cap', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_cap + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit[c_n].flatten() for c_n in col_names_hyp], axis=1) #adjustment terms if pystan_ver == 2: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB']), axis=1) else: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB'].T), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid_sp @ cells_ca_mu cells_LcA_med = celldist_valid_sp @ cells_ca_med cells_LcA_sig = np.sqrt(celldist_valid_sp.power(2) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
19,298
46.185819
130
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/pystan/regression_pystan_model1_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 13 18:22:15 2021 @author: glavrent """ #imprort libraries import os import pathlib import glob import re #regular expression package import pickle from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') def RunStan(df_flatfile, stan_model_fname, out_fname, out_dir, res_name='res', runstan_flag=True, n_iter=600, n_chains=4, adapt_delta=0.8, max_treedepth=10, pystan_ver=2, pystan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, and a spatially independent site constant. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. runstan_flag : bool, optional Flag for running stan. If true run regression, if false read past regression output and summarize non-ergodic parameters. The default is True. n_iter : integer, optional Number of stan samples. The default is 600. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Read Data #============================ #read stan model with open(stan_model_fname, "r") as f: stan_model_code = f.read() ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']].values, axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #verify no collocated events eq_dist_min = np.min([np.linalg.norm(x_eq - np.delete(X_eq,k, axis=0), axis=1).min() for k, x_eq in enumerate(X_eq) ]) assert(eq_dist_min > 5e-5),'Error. Singular covariance matrix due to collocated events' #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #verify no collocated stations sta_dist_min = np.min([np.linalg.norm(x_sta - np.delete(X_sta,k, axis=0), axis=1).min() for k, x_sta in enumerate(X_sta) ]) assert(sta_dist_min > 5e-5),'Error. Singular covariance matrix due to collocated stations' #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #stan data stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'rec_mu': np.zeros(y_data.shape), 'Y': y_data, } stan_data_fname = out_fname + '_stan_data' + '.Rdata' ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #filename for STAN regression raw output file saved as pkl stan_fit_fname = out_dir + out_fname + '_stan_fit' + '.pkl' #run stan if runstan_flag: #control paramters control_stan = {'adapt_delta':adapt_delta, 'max_treedepth':max_treedepth} if pystan_ver == 2: import pystan if (not pystan_parallel) or n_cpu<=n_chains: #compile stan_model = pystan.StanModel(model_code=stan_model_code) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=10, control = control_stan) else: #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #multi-processing arguments os.environ['STAN_NUM_THREADS'] = str(n_cpu_chain) extra_compile_args = ['-pthread', '-DSTAN_THREADS'] #compile stan_model = pystan.StanModel(model_code=stan_model_code, extra_compile_args=extra_compile_args) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=1, control = control_stan) elif pystan_ver == 3: import nest_asyncio import stan nest_asyncio.apply() #compile stan_model = stan.build(stan_model_code, data=stan_data, random_seed=1) #full Bayesian statistics stan_fit = stan_model.sample(num_chains=n_chains, num_samples=n_iter, max_depth=max_treedepth, delta=adapt_delta) #save stan model and fit pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) with open(stan_fit_fname, "wb") as f: pickle.dump({'model' : stan_model, 'fit' : stan_fit}, f, protocol=-1) else: #load model and fit for postprocessing if has already been executed with open(stan_fit_fname, "rb") as f: data_dict = pickle.load(f) stan_fit = data_dict['fit'] stan_model = data_dict['model'] del data_dict ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit[c_n].flatten() for c_n in col_names_hyp], axis=1) #adjustment terms if pystan_ver == 2: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB']), axis=1) else: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB'].T), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summarize non-ergodic coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction and residuals #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summarize predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
15,691
44.616279
130
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/pystan/regression_pystan_model2_uncorr_cells_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 13 18:22:15 2021 @author: glavrent """ #load variables import os import pathlib import glob import re #regular expression package import pickle from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_a_erg=0, runstan_flag=True, n_iter=600, n_chains=4, adapt_delta=0.8, max_treedepth=10, pystan_ver=2, pystan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and uncorrelated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter : integer, optional Number of stan samples. The default is 600. n_chains : integer, optional Number of MCMC chains. The default is 4. runstan_flag : bool, optional Flag for running stan. If true run regression, if false read past regression output and summarize non-ergodic parameters. The default is True. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Read Data #============================ #read stan model with open(stan_model_fname, "r") as f: stan_model_code = f.read() ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'rec_mu': np.zeros(y_data.shape), 'RC': celldist_valid.to_numpy(), 'c_a_erg': c_a_erg, 'Y': y_data, } stan_data_fname = out_fname + '_stan_data' + '.Rdata' ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #filename for STAN regression raw output file saved as pkl stan_fit_fname = out_dir + out_fname + '_stan_fit' + '.pkl' #run stan if runstan_flag: #control paramters control_stan = {'adapt_delta':adapt_delta, 'max_treedepth':max_treedepth} if pystan_ver == 2: import pystan if (not pystan_parallel) or n_cpu<=n_chains: #compile stan_model = pystan.StanModel(model_code=stan_model_code) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=10, control = control_stan) else: #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #multi-processing arguments os.environ['STAN_NUM_THREADS'] = str(n_cpu_chain) extra_compile_args = ['-pthread', '-DSTAN_THREADS'] #compile stan_model = pystan.StanModel(model_code=stan_model_code, extra_compile_args=extra_compile_args) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=1, control = control_stan) elif pystan_ver == 3: import nest_asyncio import stan nest_asyncio.apply() #compile stan_model = stan.build(stan_model_code, data=stan_data, random_seed=1) #full Bayesian statistics stan_fit = stan_model.sample(num_chains=n_chains, num_samples=n_iter, max_depth=max_treedepth, delta=adapt_delta) #save stan model and fit pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) with open(stan_fit_fname, "wb") as f: pickle.dump({'model' : stan_model, 'fit' : stan_fit}, f, protocol=-1) else: #load model and fit for postprocessing if has already been executed with open(stan_fit_fname, "rb") as f: data_dict = pickle.load(f) stan_fit = data_dict['fit'] stan_model = data_dict['model'] del data_dict ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'mu_cap', 'omega_cap', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_cap + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit[c_n].flatten() for c_n in col_names_hyp], axis=1) #adjustment terms if pystan_ver == 2: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB']), axis=1) else: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB'].T), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid.values @ cells_ca_mu cells_LcA_med = celldist_valid.values @ cells_ca_med cells_LcA_sig = np.sqrt(np.square(celldist_valid.values) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
19,042
46.136139
130
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/pystan/regression_pystan_model2_corr_cells_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 13 18:22:15 2021 @author: glavrent """ #load variables import os import pathlib import glob import re #regular expression package import pickle from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_a_erg=0, runstan_flag=True, n_iter=600, n_chains=4, adapt_delta=0.8, max_treedepth=10, pystan_ver=2, pystan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and partially spatially correlated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter : integer, optional Number of stan samples. The default is 600. n_chains : integer, optional Number of MCMC chains. The default is 4. runstan_flag : bool, optional Flag for running stan. If true run regression, if false read past regression output and summarize non-ergodic parameters. The default is True. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Read Data #============================ #read stan model with open(stan_model_fname, "r") as f: stan_model_code = f.read() ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'rec_mu': np.zeros(y_data.shape), 'RC': celldist_valid.to_numpy(), 'c_a_erg': c_a_erg, 'Y': y_data, } stan_data_fname = out_fname + '_stan_data' + '.Rdata' ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #filename for STAN regression raw output file saved as pkl stan_fit_fname = out_dir + out_fname + '_stan_fit' + '.pkl' #run stan if runstan_flag: #control paramters control_stan = {'adapt_delta':adapt_delta, 'max_treedepth':max_treedepth} if pystan_ver == 2: import pystan if (not pystan_parallel) or n_cpu<=n_chains: #compile stan_model = pystan.StanModel(model_code=stan_model_code) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=10, control = control_stan) else: #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #multi-processing arguments os.environ['STAN_NUM_THREADS'] = str(n_cpu_chain) extra_compile_args = ['-pthread', '-DSTAN_THREADS'] #compile stan_model = pystan.StanModel(model_code=stan_model_code, extra_compile_args=extra_compile_args) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=1, control = control_stan) elif pystan_ver == 3: import nest_asyncio import stan nest_asyncio.apply() #compile stan_model = stan.build(stan_model_code, data=stan_data, random_seed=1) #full Bayesian statistics stan_fit = stan_model.sample(num_chains=n_chains, num_samples=n_iter, max_depth=max_treedepth, delta=adapt_delta) #save stan model and fit pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) with open(stan_fit_fname, "wb") as f: pickle.dump({'model' : stan_model, 'fit' : stan_fit}, f, protocol=-1) else: #load model and fit for postprocessing if has already been executed with open(stan_fit_fname, "rb") as f: data_dict = pickle.load(f) stan_fit = data_dict['fit'] stan_model = data_dict['model'] del data_dict ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'mu_cap', 'ell_ca1p', 'omega_ca1p', 'omega_ca2p', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_cap + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit[c_n].flatten() for c_n in col_names_hyp], axis=1) #adjustment terms if pystan_ver == 2: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB']), axis=1) else: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB'].T), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid.values @ cells_ca_mu cells_LcA_med = celldist_valid.values @ cells_ca_med cells_LcA_sig = np.sqrt(np.square(celldist_valid.values) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
19,072
46.093827
130
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/pystan/regression_pystan_model2_corr_cells_sparse_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 13 18:22:15 2021 @author: glavrent """ #load variables import os import pathlib import glob import re #regular expression package import pickle from joblib import cpu_count #arithmetic libraries import numpy as np from scipy import sparse #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_a_erg=0, runstan_flag=True, n_iter=600, n_chains=4, adapt_delta=0.8, max_treedepth=10, pystan_ver=2, pystan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and partially spatially correlated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter : integer, optional Number of stan samples. The default is 600. n_chains : integer, optional Number of MCMC chains. The default is 4. runstan_flag : bool, optional Flag for running stan. If true run regression, if false read past regression output and summarize non-ergodic parameters. The default is True. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Read Data #============================ #read stan model with open(stan_model_fname, "r") as f: stan_model_code = f.read() ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid].to_numpy() #cell-distance with only non-zero cells celldist_valid_sp = sparse.csr_matrix(celldist_valid) #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', np.abs(df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'NCELL_SP': len(celldist_valid_sp.data), 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'rec_mu': np.zeros(y_data.shape), 'RC_val': celldist_valid_sp.data, 'RC_w': celldist_valid_sp.indices+1, 'RC_u': celldist_valid_sp.indptr+1, 'c_a_erg': c_a_erg, 'Y': y_data, } stan_data_fname = out_fname + '_stan_data' + '.Rdata' ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #filename for STAN regression raw output file saved as pkl stan_fit_fname = out_dir + out_fname + '_stan_fit' + '.pkl' #run stan if runstan_flag: #control paramters control_stan = {'adapt_delta':adapt_delta, 'max_treedepth':max_treedepth} if pystan_ver == 2: import pystan if (not pystan_parallel) or n_cpu<=n_chains: #compile stan_model = pystan.StanModel(model_code=stan_model_code) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=10, control = control_stan) else: #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #multi-processing arguments os.environ['STAN_NUM_THREADS'] = str(n_cpu_chain) extra_compile_args = ['-pthread', '-DSTAN_THREADS'] #compile stan_model = pystan.StanModel(model_code=stan_model_code, extra_compile_args=extra_compile_args) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=1, control = control_stan) elif pystan_ver == 3: import nest_asyncio import stan nest_asyncio.apply() #compile stan_model = stan.build(stan_model_code, data=stan_data, random_seed=1) #full Bayesian statistics stan_fit = stan_model.sample(num_chains=n_chains, num_samples=n_iter, max_depth=max_treedepth, delta=adapt_delta) #save stan model and fit pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) with open(stan_fit_fname, "wb") as f: pickle.dump({'model' : stan_model, 'fit' : stan_fit}, f, protocol=-1) else: #load model and fit for postprocessing if has already been executed with open(stan_fit_fname, "rb") as f: data_dict = pickle.load(f) stan_fit = data_dict['fit'] stan_model = data_dict['model'] del data_dict ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'mu_cap', 'ell_ca1p', 'omega_ca1p', 'omega_ca2p', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_cap + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit[c_n].flatten() for c_n in col_names_hyp], axis=1) #adjustment terms if pystan_ver == 2: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB']), axis=1) else: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB'].T), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid_sp @ cells_ca_mu cells_LcA_med = celldist_valid_sp @ cells_ca_med cells_LcA_sig = np.sqrt(celldist_valid_sp.power(2) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
19,326
46.139024
130
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/pystan/regression_pystan_model3_corr_cells_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 13 18:22:15 2021 @author: glavrent """ #load variables import os import pathlib import glob import re #regular expression package import pickle from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_2_erg=0, c_3_erg=0, c_a_erg=0, runstan_flag=True, n_iter=600, n_chains=4, adapt_delta=0.8, max_treedepth=10, pystan_ver=2, pystan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and partially spatially correlated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_2_erg : double, optional Value of ergodic geometrical spreading coefficient. The default is 0. c_3_erg : double, optional Value of ergodic Vs30 coefficient. The default is 0. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter : integer, optional Number of stan samples. The default is 600. n_chains : integer, optional Number of MCMC chains. The default is 4. runstan_flag : bool, optional Flag for running stan. If true run regression, if false read past regression output and summarize non-ergodic parameters. The default is True. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' #number of cores n_cpu = max(cpu_count() -1,1) ## Read Data #============================ #read stan model with open(stan_model_fname, "r") as f: stan_model_code = f.read() ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','x_3','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[3,4]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #geometrical spreading covariates x_2 = df_flatfile['x_2'].values #vs30 covariates x_3 = df_flatfile['x_3'].values[sta_idx] #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'rec_mu': np.zeros(y_data.shape), 'Y': y_data, 'x_2': x_2, 'x_3': x_3, 'c_2_erg': c_2_erg, 'c_3_erg': c_3_erg, 'c_a_erg': c_a_erg, 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'RC': celldist_valid.to_numpy(), } stan_data_fname = out_fname + '_stan_data' + '.Rdata' ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #filename for STAN regression raw output file saved as pkl stan_fit_fname = out_dir + out_fname + '_stan_fit' + '.pkl' #run stan if runstan_flag: #control paramters control_stan = {'adapt_delta':adapt_delta, 'max_treedepth':max_treedepth} if pystan_ver == 2: import pystan if (not pystan_parallel) or n_cpu<=n_chains: #compile stan_model = pystan.StanModel(model_code=stan_model_code) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=10, control = control_stan) else: #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #multi-processing arguments os.environ['STAN_NUM_THREADS'] = str(n_cpu_chain) extra_compile_args = ['-pthread', '-DSTAN_THREADS'] #compile stan_model = pystan.StanModel(model_code=stan_model_code, extra_compile_args=extra_compile_args) #full Bayesian statistics stan_fit = stan_model.sampling(data=stan_data, iter=n_iter, chains = n_chains, refresh=1, control = control_stan) elif pystan_ver == 3: import nest_asyncio import stan nest_asyncio.apply() #compile stan_model = stan.build(stan_model_code, data=stan_data, random_seed=1) #full Bayesian statistics stan_fit = stan_model.sample(num_chains=n_chains, num_samples=n_iter, max_depth=max_treedepth, delta=adapt_delta) #save stan model and fit pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) with open(stan_fit_fname, "wb") as f: pickle.dump({'model' : stan_model, 'fit' : stan_fit}, f, protocol=-1) else: #load model and fit for postprocessing if has already been executed with open(stan_fit_fname, "rb") as f: data_dict = pickle.load(f) stan_fit = data_dict['fit'] stan_model = data_dict['model'] del data_dict ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','mu_2p','mu_3s', 'ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'ell_2p', 'ell_3s', 'omega_2p', 'omega_3s', 'mu_cap', 'ell_ca1p', 'omega_ca1p', 'omega_ca2p', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_c_2p = ['c_2p.%i'%(k) for k in range(n_eq)] col_names_c_3s = ['c_3s.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = (col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_c_2p + col_names_c_3s + col_names_cap + col_names_dB) #summarize raw posterior distributions stan_posterior = np.stack([stan_fit[c_n].flatten() for c_n in col_names_hyp], axis=1) #adjustment terms if pystan_ver == 2: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_2p']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_3s']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap']), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB']), axis=1) else: stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1e'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1as'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dc_1bs'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_2p'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_3s'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['c_cap'].T), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit['dB'].T), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_c_2p, col_names_c_3s, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid.values @ cells_ca_mu cells_LcA_med = celldist_valid.values @ cells_ca_med cells_LcA_sig = np.sqrt(np.square(celldist_valid.values) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] #spatially varying geometrical spreading coefficient coeff_2p_mu = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].mean() for k in range(n_eq)]) coeff_2p_mu = coeff_2p_mu[eq_inv] coeff_2p_med = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].median() for k in range(n_eq)]) coeff_2p_med = coeff_2p_med[eq_inv] coeff_2p_sig = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].std() for k in range(n_eq)]) coeff_2p_sig = coeff_2p_sig[eq_inv] #spatially varying Vs30 coefficient coeff_3s_mu = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].mean() for k in range(n_sta)]) coeff_3s_mu = coeff_3s_mu[sta_inv] coeff_3s_med = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].median() for k in range(n_sta)]) coeff_3s_med = coeff_3s_med[sta_inv] coeff_3s_sig = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].std() for k in range(n_sta)]) coeff_3s_sig = coeff_3s_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, coeff_2p_mu, coeff_3s_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, coeff_2p_med, coeff_3s_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, coeff_2p_sig, coeff_3s_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','c_2p_mean','c_3s_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'c_2p_med', 'c_3s_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'c_2p_sig', 'c_3s_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + coeff_2p_mu*x_2 + coeff_3s_mu*x_3[sta_inv] + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
21,689
46.986726
130
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model1_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Jul 13 18:22:15 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, stan_model_fname, out_fname, out_dir, res_name='res', n_iter_warmup=300, n_iter_sampling=300, n_chains=4, max_treedepth=10, adapt_delta=0.80, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, and a spatially independent site constant. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. stan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']].values, axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #verify no collocated events eq_dist_min = np.min([np.linalg.norm(x_eq - np.delete(X_eq,k, axis=0), axis=1).min() for k, x_eq in enumerate(X_eq) ]) assert(eq_dist_min > 5e-5),'Error. Singular covariance matrix due to collocated events' #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #verify no collocated stations sta_dist_min = np.min([np.linalg.norm(x_sta - np.delete(X_sta,k, axis=0), axis=1).min() for k, x_sta in enumerate(X_sta) ]) assert(sta_dist_min > 5e-5),'Error. Singular covariance matrix due to collocated stations' #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #stan data stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'rec_mu': np.zeros(y_data.shape), 'Y': y_data, } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior.to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit.summary(), file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
14,382
45.247588
127
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model2_corr_cells_sparse_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:13:49 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np from scipy import sparse #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_a_erg=0, n_iter_warmup=300, n_iter_sampling=300, n_chains=4, adapt_delta=0.8, max_treedepth=10, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and partially spatially correlated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. stan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid].to_numpy() #cell-distance with only non-zero cells celldist_valid_sp = sparse.csr_matrix(celldist_valid) #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', np.abs(df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'NCELL_SP': len(celldist_valid_sp.data), 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'rec_mu': np.zeros(y_data.shape), 'RC_val': celldist_valid_sp.data, 'RC_w': celldist_valid_sp.indices+1, 'RC_u': celldist_valid_sp.indptr+1, 'c_a_erg': c_a_erg, 'Y': y_data, } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'mu_cap', 'ell_ca1p', 'omega_ca1p', 'omega_ca2p', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_cap + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_cap')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior.to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid_sp @ cells_ca_mu cells_LcA_med = celldist_valid_sp @ cells_ca_med cells_LcA_sig = np.sqrt(celldist_valid_sp.power(2) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #dataframe with flatfile info df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
17,927
46.808
120
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model2_uncorr_cells_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:13:49 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_a_erg=0, n_iter_warmup=300, n_iter_sampling=300, n_chains=4, adapt_delta=0.8, max_treedepth=10, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and uncorrelated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'rec_mu': np.zeros(y_data.shape), 'RC': celldist_valid.to_numpy(), 'c_a_erg': c_a_erg, 'Y': y_data, } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'mu_cap', 'omega_cap', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_cap + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_cap')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid.values @ cells_ca_mu cells_LcA_med = celldist_valid.values @ cells_ca_med cells_LcA_sig = np.sqrt(np.square(celldist_valid.values) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
17,759
46.741935
120
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model3_uncorr_cells_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:13:49 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_2_erg=0, c_3_erg=0, c_a_erg=0, n_iter_warmup=300, n_iter_sampling=300, n_chains=4, adapt_delta=0.8, max_treedepth=10, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and uncorrelated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','x_3','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[3,4]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #geometrical spreading covariates x_2 = df_flatfile['x_2'].values #vs30 covariates x_3 = df_flatfile['x_3'].values[sta_idx] #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'rec_mu': np.zeros(y_data.shape), 'Y': y_data, 'x_2': x_2, 'x_3': x_3, 'c_2_erg': c_2_erg, 'c_3_erg': c_3_erg, 'c_a_erg': c_a_erg, 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'RC': celldist_valid.to_numpy(), } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','mu_2p','mu_3s', 'ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'ell_2p', 'ell_3s', 'omega_2p', 'omega_3s', 'mu_cap', 'omega_cap', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_c_2p = ['c_2p.%i'%(k) for k in range(n_eq)] col_names_c_3s = ['c_3s.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = (col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_c_2p + col_names_c_3s + col_names_cap + col_names_dB) #sumarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_2p')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_3s')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_cap')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_c_2p, col_names_c_3s, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid.values @ cells_ca_mu cells_LcA_med = celldist_valid.values @ cells_ca_med cells_LcA_sig = np.sqrt(np.square(celldist_valid.values) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] #spatially varying geometrical spreading coefficient coeff_2p_mu = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].mean() for k in range(n_eq)]) coeff_2p_mu = coeff_2p_mu[eq_inv] coeff_2p_med = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].median() for k in range(n_eq)]) coeff_2p_med = coeff_2p_med[eq_inv] coeff_2p_sig = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].std() for k in range(n_eq)]) coeff_2p_sig = coeff_2p_sig[eq_inv] #spatially varying Vs30 coefficient coeff_3s_mu = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].mean() for k in range(n_sta)]) coeff_3s_mu = coeff_3s_mu[sta_inv] coeff_3s_med = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].median() for k in range(n_sta)]) coeff_3s_med = coeff_3s_med[sta_inv] coeff_3s_sig = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].std() for k in range(n_sta)]) coeff_3s_sig = coeff_3s_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #dataframe with flatfile info df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, coeff_2p_mu, coeff_3s_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, coeff_2p_med, coeff_3s_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, coeff_2p_sig, coeff_3s_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','c_2p_mean','c_3s_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'c_2p_med', 'c_3s_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'c_2p_sig', 'c_3s_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + coeff_2p_mu*x_2 + coeff_3s_mu*x_3[sta_inv] + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
19,904
47.667482
128
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model3_uncorr_cells_sparse_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:13:49 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np from scipy import sparse #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_2_erg=0, c_3_erg=0, c_a_erg=0, n_iter_warmup=300, n_iter_sampling=300, n_chains=4, adapt_delta=0.8, max_treedepth=10, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and uncorrelated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','x_3','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[3,4]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #geometrical spreading covariates x_2 = df_flatfile['x_2'].values #vs30 covariates x_3 = df_flatfile['x_3'].values[sta_idx] #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells celldist_valid_sp = sparse.csr_matrix(celldist_valid) #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'NCELL_SP': len(celldist_valid_sp.data), 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'rec_mu': np.zeros(y_data.shape), 'Y': y_data, 'x_2': x_2, 'x_3': x_3, 'c_2_erg': c_2_erg, 'c_3_erg': c_3_erg, 'c_a_erg': c_a_erg, 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'RC_val': celldist_valid_sp.data, 'RC_w': celldist_valid_sp.indices+1, 'RC_u': celldist_valid_sp.indptr+1, } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','mu_2p','mu_3s', 'ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'ell_2p', 'ell_3s', 'omega_2p', 'omega_3s', 'mu_cap', 'omega_cap', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_c_2p = ['c_2p.%i'%(k) for k in range(n_eq)] col_names_c_3s = ['c_3s.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = (col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_c_2p + col_names_c_3s + col_names_cap + col_names_dB) #sumarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_2p')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_3s')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_cap')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior.to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_c_2p, col_names_c_3s, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid_sp @ cells_ca_mu cells_LcA_med = celldist_valid_sp @ cells_ca_med cells_LcA_sig = np.sqrt(celldist_valid_sp.power(2) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] #spatially varying geometrical spreading coefficient coeff_2p_mu = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].mean() for k in range(n_eq)]) coeff_2p_mu = coeff_2p_mu[eq_inv] coeff_2p_med = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].median() for k in range(n_eq)]) coeff_2p_med = coeff_2p_med[eq_inv] coeff_2p_sig = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].std() for k in range(n_eq)]) coeff_2p_sig = coeff_2p_sig[eq_inv] #spatially varying Vs30 coefficient coeff_3s_mu = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].mean() for k in range(n_sta)]) coeff_3s_mu = coeff_3s_mu[sta_inv] coeff_3s_med = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].median() for k in range(n_sta)]) coeff_3s_med = coeff_3s_med[sta_inv] coeff_3s_sig = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].std() for k in range(n_sta)]) coeff_3s_sig = coeff_3s_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #dataframe with flatfile info df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, coeff_2p_mu, coeff_3s_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, coeff_2p_med, coeff_3s_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, coeff_2p_sig, coeff_3s_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','c_2p_mean','c_3s_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'c_2p_med', 'c_3s_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'c_2p_sig', 'c_3s_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + coeff_2p_mu*x_2 + coeff_3s_mu*x_3[sta_inv] + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
20,144
47.65942
128
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model2_corr_cells_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:13:49 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_a_erg=0, n_iter_warmup=300, n_iter_sampling=300, n_chains=4, adapt_delta=0.8, max_treedepth=10, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and partially spatially correlated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. stan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'rec_mu': np.zeros(y_data.shape), 'RC': celldist_valid.to_numpy(), 'c_a_erg': c_a_erg, 'Y': y_data, } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'mu_cap', 'ell_ca1p', 'omega_ca1p', 'omega_ca2p', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_cap + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_cap')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior.to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid.values @ cells_ca_mu cells_LcA_med = celldist_valid.values @ cells_ca_med cells_LcA_sig = np.sqrt(np.square(celldist_valid.values) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #dataframe with flatfile info df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
17,672
46.764865
120
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model3_corr_cells_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:13:49 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_2_erg=0, c_3_erg=0, c_a_erg=0, n_iter_warmup=300, n_iter_sampling=300, n_chains=4, adapt_delta=0.8, max_treedepth=10, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and partially spatially correlated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_2_erg : double, optional Value of ergodic geometrical spreading coefficient. The default is 0. c_3_erg : double, optional Value of ergodic Vs30 coefficient. The default is 0. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. stan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','x_3','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[3,4]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #geometrical spreading covariates x_2 = df_flatfile['x_2'].values #vs30 covariates x_3 = df_flatfile['x_3'].values[sta_idx] #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'rec_mu': np.zeros(y_data.shape), 'Y': y_data, 'x_2': x_2, 'x_3': x_3, 'c_2_erg': c_2_erg, 'c_3_erg': c_3_erg, 'c_a_erg': c_a_erg, 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'RC': celldist_valid.to_numpy(), } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','mu_2p','mu_3s', 'ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'ell_2p', 'ell_3s', 'omega_2p', 'omega_3s', 'mu_cap', 'ell_ca1p', 'omega_ca1p', 'omega_ca2p', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_c_2p = ['c_2p.%i'%(k) for k in range(n_eq)] col_names_c_3s = ['c_3s.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = (col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_c_2p + col_names_c_3s + col_names_cap + col_names_dB) #summarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_2p')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_3s')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_cap')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior .to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_c_2p, col_names_c_3s, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid.values @ cells_ca_mu cells_LcA_med = celldist_valid.values @ cells_ca_med cells_LcA_sig = np.sqrt(np.square(celldist_valid.values) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] #spatially varying geometrical spreading coefficient coeff_2p_mu = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].mean() for k in range(n_eq)]) coeff_2p_mu = coeff_2p_mu[eq_inv] coeff_2p_med = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].median() for k in range(n_eq)]) coeff_2p_med = coeff_2p_med[eq_inv] coeff_2p_sig = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].std() for k in range(n_eq)]) coeff_2p_sig = coeff_2p_sig[eq_inv] #spatially varying Vs30 coefficient coeff_3s_mu = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].mean() for k in range(n_sta)]) coeff_3s_mu = coeff_3s_mu[sta_inv] coeff_3s_med = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].median() for k in range(n_sta)]) coeff_3s_med = coeff_3s_med[sta_inv] coeff_3s_sig = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].std() for k in range(n_sta)]) coeff_3s_sig = coeff_3s_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, coeff_2p_mu, coeff_3s_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, coeff_2p_med, coeff_3s_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, coeff_2p_sig, coeff_3s_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','c_2p_mean','c_3s_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'c_2p_med', 'c_3s_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'c_2p_sig', 'c_3s_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + coeff_2p_mu*x_2 + coeff_3s_mu*x_3[sta_inv] + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
20,112
47.699758
128
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model2_uncorr_cells_sparse_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:13:49 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np from scipy import sparse #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_a_erg=0, n_iter_warmup=300, n_iter_sampling=300, n_chains=4, adapt_delta=0.8, max_treedepth=10, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and uncorrelated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. pystan_ver : integer, optional Version of pystan to run. The default is 2. pystan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid].to_numpy() #cell-distance with only non-zero cells celldist_valid_sp = sparse.csr_matrix(celldist_valid) #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', np.abs(df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'NCELL_SP': len(celldist_valid_sp.data), 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'rec_mu': np.zeros(y_data.shape), 'RC_val': celldist_valid_sp.data, 'RC_w': celldist_valid_sp.indices+1, 'RC_u': celldist_valid_sp.indptr+1, 'c_a_erg': c_a_erg, 'Y': y_data, } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'mu_cap', 'omega_cap', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_cap + col_names_dB #summarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_cap')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior.to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid_sp @ cells_ca_mu cells_LcA_med = celldist_valid_sp @ cells_ca_med cells_LcA_sig = np.sqrt(celldist_valid_sp.power(2) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
18,013
46.782493
120
py
ngmm_tools
ngmm_tools-master/Analyses/Python_lib/regression/cmdstan/regression_cmdstan_model3_corr_cells_sparse_unbounded_hyp.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:13:49 2021 @author: glavrent """ #load variables import pathlib from joblib import cpu_count #arithmetic libraries import numpy as np from scipy import sparse #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl from matplotlib.ticker import AutoLocator as plt_autotick import arviz as az mpl.use('agg') #stan library import cmdstanpy def RunStan(df_flatfile, df_cellinfo, df_celldist, stan_model_fname, out_fname, out_dir, res_name='res', c_2_erg=0, c_3_erg=0, c_a_erg=0, n_iter_warmup=300, n_iter_sampling=300, n_chains=4, adapt_delta=0.8, max_treedepth=10, stan_parallel=False): ''' Run full Bayessian regression in Stan. Non-ergodic model includes: a spatially varying earthquake constant, a spatially varying site constant, a spatially independent site constant, and partially spatially correlated anelastic attenuation. Parameters ---------- df_flatfile : pd.DataFrame Input data frame containing total residuals, eq and site coordinates. df_cellinfo : pd.DataFrame Dataframe with coordinates of anelastic attenuation cells. df_celldist : pd.DataFrame Datafame with cell path distances of all records in df_flatfile. stan_model_fname : string File name for stan model. out_fname : string File name for output files. out_dir : string Output directory. res_name : string, optional Column name for total residuals. The default is 'res'. c_2_erg : double, optional Value of ergodic geometrical spreading coefficient. The default is 0. c_3_erg : double, optional Value of ergodic Vs30 coefficient. The default is 0. c_a_erg : double, optional Value of ergodic anelatic attenuation coefficient. Used as mean of cell specific anelastic attenuation prior distribution. The default is 0. n_iter_warmup : integer, optional Number of burn out MCMC samples. The default is 300. n_iter_sampling : integer, optional Number of MCMC samples for computing the posterior distributions. The default is 300. n_chains : integer, optional Number of MCMC chains. The default is 4. adapt_delta : double, optional Target average proposal acceptance probability for adaptation. The default is 0.8. max_treedepth : integer, optional Maximum number of evaluations for each iteration (2^max_treedepth). The default is 10. stan_parallel : bool, optional Flag for using multithreaded option in STAN. The default is False. Returns ------- None. ''' ## Preprocess Input Data #============================ #set rsn column as dataframe index, skip if rsn already the index if not df_flatfile.index.name == 'rsn': df_flatfile.set_index('rsn', drop=True, inplace=True) if not df_celldist.index.name == 'rsn': df_celldist.set_index('rsn', drop=True, inplace=True) #set cellid column as dataframe index, skip if cellid already the index if not df_cellinfo.index.name == 'cellid': df_cellinfo.set_index('cellid', drop=True, inplace=True) #number of data n_data = len(df_flatfile) #earthquake data data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_inverse=True, return_index=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates #create earthquake ids for all records (1 to n_eq) eq_id = eq_inv + 1 n_eq = len(data_eq) #station data data_sta_all = df_flatfile[['ssn','Vs30','x_3','staX','staY']].values _, sta_idx, sta_inv = np.unique( df_flatfile[['ssn']].values, axis = 0, return_inverse=True, return_index=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[3,4]] #station coordinates #create station indices for all records (1 to n_sta) sta_id = sta_inv + 1 n_sta = len(data_sta) #geometrical spreading covariates x_2 = df_flatfile['x_2'].values #vs30 covariates x_3 = df_flatfile['x_3'].values[sta_idx] #ground-motion observations y_data = df_flatfile[res_name].to_numpy().copy() #cell data #reorder and only keep records included in the flatfile df_celldist = df_celldist.reindex(df_flatfile.index) #cell info cell_ids_all = df_cellinfo.index cell_names_all = df_cellinfo.cellname #cell distance matrix celldist_all = df_celldist[cell_names_all] #cell-distance matrix with all cells #find cell with more than one paths i_cells_valid = np.where(celldist_all.sum(axis=0) > 0)[0] #valid cells with more than one path cell_ids_valid = cell_ids_all[i_cells_valid] cell_names_valid = cell_names_all[i_cells_valid] celldist_valid = celldist_all.loc[:,cell_names_valid] #cell-distance with only non-zero cells celldist_valid_sp = sparse.csr_matrix(celldist_valid) #number of cells n_cell = celldist_all.shape[1] n_cell_valid = celldist_valid.shape[1] #cell coordinates X_cells_valid = df_cellinfo.loc[i_cells_valid,['mptX','mptY']].values #print Rrup missfits print('max R_rup misfit', (df_flatfile.Rrup.values - celldist_valid.sum(axis=1)).abs().max()) stan_data = {'N': n_data, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell_valid, 'NCELL_SP': len(celldist_valid_sp.data), 'eq': eq_id, #earthquake id 'stat': sta_id, #station id 'rec_mu': np.zeros(y_data.shape), 'Y': y_data, 'x_2': x_2, 'x_3': x_3, 'c_2_erg': c_2_erg, 'c_3_erg': c_3_erg, 'c_a_erg': c_a_erg, 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cells_valid, 'RC_val': celldist_valid_sp.data, 'RC_w': celldist_valid_sp.indices+1, 'RC_u': celldist_valid_sp.indptr+1, } stan_data_fname = out_dir + out_fname + '_stan_data' + '.json' #create output directory pathlib.Path(out_dir).mkdir(parents=True, exist_ok=True) #write as json file cmdstanpy.utils.write_stan_json(stan_data_fname, stan_data) ## Run Stan, fit model #============================ #number of cores n_cpu = max(cpu_count() -1,1) #run stan if (not stan_parallel) or n_cpu<=n_chains: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname) stan_model.compile(force=True) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') else: #compile stan model stan_model = cmdstanpy.CmdStanModel(stan_file=stan_model_fname, cpp_options={"STAN_THREADS": True}) stan_model.compile(force=True) #number of cores per chain n_cpu_chain = int(np.floor(n_cpu/n_chains)) #run full MCMC sampler stan_fit = stan_model.sample(data=stan_data_fname, chains=n_chains, threads_per_chain=n_cpu_chain, iter_warmup=n_iter_warmup, iter_sampling=n_iter_sampling, seed=1, max_treedepth=max_treedepth, adapt_delta=adapt_delta, show_progress=True, output_dir=out_dir+'stan_fit/') ## Postprocessing Data #============================ ## Extract posterior samples # --------------------------- #hyper-parameters col_names_hyp = ['dc_0','mu_2p','mu_3s', 'ell_1e', 'ell_1as', 'omega_1e', 'omega_1as', 'omega_1bs', 'ell_2p', 'ell_3s', 'omega_2p', 'omega_3s', 'mu_cap', 'ell_ca1p', 'omega_ca1p', 'omega_ca2p', 'phi_0','tau_0'] #non-ergodic terms col_names_dc_1e = ['dc_1e.%i'%(k) for k in range(n_eq)] col_names_dc_1as = ['dc_1as.%i'%(k) for k in range(n_sta)] col_names_dc_1bs = ['dc_1bs.%i'%(k) for k in range(n_sta)] col_names_c_2p = ['c_2p.%i'%(k) for k in range(n_eq)] col_names_c_3s = ['c_3s.%i'%(k) for k in range(n_sta)] col_names_dB = ['dB.%i'%(k) for k in range(n_eq)] col_names_cap = ['c_cap.%i'%(c_id) for c_id in cell_ids_valid] col_names_all = (col_names_hyp + col_names_dc_1e + col_names_dc_1as + col_names_dc_1bs + col_names_c_2p + col_names_c_3s + col_names_cap + col_names_dB) #summarize raw posterior distributions stan_posterior = np.stack([stan_fit.stan_variable(c_n) for c_n in col_names_hyp], axis=1) #adjustment terms stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1e')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1as')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dc_1bs')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_2p')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_3s')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('c_cap')), axis=1) stan_posterior = np.concatenate((stan_posterior, stan_fit.stan_variable('dB')), axis=1) #save raw-posterior distribution df_stan_posterior_raw = pd.DataFrame(stan_posterior, columns = col_names_all) df_stan_posterior_raw.to_csv(out_dir + out_fname + '_stan_posterior_raw' + '.csv', index=False) ## Summarize hyper-parameters # --------------------------- #summarize posterior distributions of hyper-parameters perc_array = np.array([0.05,0.25,0.5,0.75,0.95]) df_stan_hyp = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_hyp = df_stan_hyp.append(df_stan_posterior_raw[col_names_hyp].mean(axis = 0), ignore_index=True) df_stan_hyp.index = ['prc_%.2f'%(prc) for prc in perc_array]+['mean'] df_stan_hyp.to_csv(out_dir + out_fname + '_stan_hyperparameters' + '.csv', index=True) #detailed posterior percentiles of posterior distributions perc_array = np.arange(0.01,0.99,0.01) df_stan_posterior = df_stan_posterior_raw[col_names_hyp].quantile(perc_array) df_stan_posterior.index.name = 'prc' df_stan_posterior.to_csv(out_dir + out_fname + '_stan_hyperposterior' + '.csv', index=True) del col_names_dc_1e, col_names_dc_1as, col_names_dc_1bs, col_names_c_2p, col_names_c_3s, col_names_dB del stan_posterior, col_names_all ## Sample spatially varying coefficients and predictions at record locations # --------------------------- # earthquake and station location in database X_eq_all = df_flatfile[['eqX', 'eqY']].values X_sta_all = df_flatfile[['staX','staY']].values # GMM anelastic attenuation #--- --- --- --- --- --- --- --- cells_ca_mu = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].mean() for k in cell_ids_valid]) cells_ca_med = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].median() for k in cell_ids_valid]) cells_ca_sig = np.array([df_stan_posterior_raw.loc[:,'c_cap.%i'%(k)].std() for k in cell_ids_valid]) #effect of anelastic attenuation in GM cells_LcA_mu = celldist_valid_sp @ cells_ca_mu cells_LcA_med = celldist_valid_sp @ cells_ca_med cells_LcA_sig = np.sqrt(celldist_valid_sp.power(2) @ cells_ca_sig**2) #summary attenuation cells catten_summary = np.vstack((np.tile(c_a_erg, n_cell_valid), cells_ca_mu, cells_ca_med, cells_ca_sig)).T columns_names = ['c_a_erg','c_cap_mean','c_cap_med','c_cap_sig'] df_catten_summary = pd.DataFrame(catten_summary, columns = columns_names, index=df_cellinfo.index[i_cells_valid]) #create dataframe with summary attenuation cells df_catten_summary = pd.merge(df_cellinfo[['cellname','mptLat','mptLon','mptX','mptY','mptZ','UTMzone']], df_catten_summary, how='right', left_index=True, right_index=True) df_catten_summary.to_csv(out_dir + out_fname + '_stan_catten' + '.csv', index=True) # GMM coefficients #--- --- --- --- --- --- --- --- #constant shift coefficient coeff_0_mu = df_stan_posterior_raw.loc[:,'dc_0'].mean() * np.ones(n_data) coeff_0_med = df_stan_posterior_raw.loc[:,'dc_0'].median() * np.ones(n_data) coeff_0_sig = df_stan_posterior_raw.loc[:,'dc_0'].std() * np.ones(n_data) #spatially varying earthquake constant coefficient coeff_1e_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].mean() for k in range(n_eq)]) coeff_1e_mu = coeff_1e_mu[eq_inv] coeff_1e_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].median() for k in range(n_eq)]) coeff_1e_med = coeff_1e_med[eq_inv] coeff_1e_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1e.{k}'].std() for k in range(n_eq)]) coeff_1e_sig = coeff_1e_sig[eq_inv] #site term constant covariance coeff_1as_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].mean() for k in range(n_sta)]) coeff_1as_mu = coeff_1as_mu[sta_inv] coeff_1as_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].median() for k in range(n_sta)]) coeff_1as_med = coeff_1as_med[sta_inv] coeff_1as_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1as.{k}'].std() for k in range(n_sta)]) coeff_1as_sig = coeff_1as_sig[sta_inv] #spatially varying station constant covariance coeff_1bs_mu = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].mean() for k in range(n_sta)]) coeff_1bs_mu = coeff_1bs_mu[sta_inv] coeff_1bs_med = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].median() for k in range(n_sta)]) coeff_1bs_med = coeff_1bs_med[sta_inv] coeff_1bs_sig = np.array([df_stan_posterior_raw.loc[:,f'dc_1bs.{k}'].std() for k in range(n_sta)]) coeff_1bs_sig = coeff_1bs_sig[sta_inv] #spatially varying geometrical spreading coefficient coeff_2p_mu = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].mean() for k in range(n_eq)]) coeff_2p_mu = coeff_2p_mu[eq_inv] coeff_2p_med = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].median() for k in range(n_eq)]) coeff_2p_med = coeff_2p_med[eq_inv] coeff_2p_sig = np.array([df_stan_posterior_raw.loc[:,f'c_2p.{k}'].std() for k in range(n_eq)]) coeff_2p_sig = coeff_2p_sig[eq_inv] #spatially varying Vs30 coefficient coeff_3s_mu = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].mean() for k in range(n_sta)]) coeff_3s_mu = coeff_3s_mu[sta_inv] coeff_3s_med = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].median() for k in range(n_sta)]) coeff_3s_med = coeff_3s_med[sta_inv] coeff_3s_sig = np.array([df_stan_posterior_raw.loc[:,f'c_3s.{k}'].std() for k in range(n_sta)]) coeff_3s_sig = coeff_3s_sig[sta_inv] # aleatory variability phi_0_array = np.array([df_stan_posterior_raw.phi_0.mean()]*X_sta_all.shape[0]) tau_0_array = np.array([df_stan_posterior_raw.tau_0.mean()]*X_sta_all.shape[0]) #initiaize flatfile for sumamry of non-erg coefficinets and residuals df_flatinfo = df_flatfile[['eqid','ssn','eqLat','eqLon','staLat','staLon','eqX','eqY','staX','staY','UTMzone']] #summary coefficients coeffs_summary = np.vstack((coeff_0_mu, coeff_1e_mu, coeff_1as_mu, coeff_1bs_mu, coeff_2p_mu, coeff_3s_mu, cells_LcA_mu, coeff_0_med, coeff_1e_med, coeff_1as_med, coeff_1bs_med, coeff_2p_med, coeff_3s_med, cells_LcA_med, coeff_0_sig, coeff_1e_sig, coeff_1as_sig, coeff_1bs_sig, coeff_2p_sig, coeff_3s_sig, cells_LcA_sig)).T columns_names = ['dc_0_mean','dc_1e_mean','dc_1as_mean','dc_1bs_mean','c_2p_mean','c_3s_mean','Lc_ca_mean', 'dc_0_med', 'dc_1e_med', 'dc_1as_med', 'dc_1bs_med', 'c_2p_med', 'c_3s_med', 'Lc_ca_med', 'dc_0_sig', 'dc_1e_sig', 'dc_1as_sig', 'dc_1bs_sig', 'c_2p_sig', 'c_3s_sig', 'Lc_ca_sig'] df_coeffs_summary = pd.DataFrame(coeffs_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with summary coefficients df_coeffs_summary = pd.merge(df_flatinfo, df_coeffs_summary, how='right', left_index=True, right_index=True) df_coeffs_summary[['eqid','ssn']] = df_coeffs_summary[['eqid','ssn']].astype(int) df_coeffs_summary.to_csv(out_dir + out_fname + '_stan_coefficients' + '.csv', index=True) # GMM prediction #--- --- --- --- --- --- --- --- #mean prediction y_mu = (coeff_0_mu + coeff_1e_mu + coeff_1as_mu + coeff_1bs_mu + coeff_2p_mu*x_2 + coeff_3s_mu*x_3[sta_inv] + cells_LcA_mu) #compute residuals res_tot = y_data - y_mu #residuals computed directly from stan regression res_between = [df_stan_posterior_raw.loc[:,f'dB.{k}'].mean() for k in range(n_eq)] res_between = np.array([res_between[k] for k in (eq_inv).astype(int)]) res_within = res_tot - res_between #summary predictions and residuals predict_summary = np.vstack((y_mu, res_tot, res_between, res_within)).T columns_names = ['nerg_mu','res_tot','res_between','res_within'] df_predict_summary = pd.DataFrame(predict_summary, columns = columns_names, index=df_flatfile.index) #create dataframe with predictions and residuals df_predict_summary = pd.merge(df_flatinfo, df_predict_summary, how='right', left_index=True, right_index=True) df_predict_summary[['eqid','ssn']] = df_predict_summary[['eqid','ssn']].astype(int) df_predict_summary.to_csv(out_dir + out_fname + '_stan_residuals' + '.csv', index=True) ## Summary regression # --------------------------- #save summary statistics stan_summary_fname = out_dir + out_fname + '_stan_summary' + '.txt' with open(stan_summary_fname, 'w') as f: print(stan_fit, file=f) #create and save trace plots fig_dir = out_dir + 'summary_figs/' #create figures directory if doesn't exit pathlib.Path(fig_dir).mkdir(parents=True, exist_ok=True) #create stan trace plots stan_az_fit = az.from_cmdstanpy(stan_fit) # stan_az_fit = az.from_cmdstanpy(stan_fit, posterior_predictive='Y') for c_name in col_names_hyp: #create trace plot with arviz ax = az.plot_trace(stan_az_fit, var_names=c_name, figsize=(10,5) ).ravel() ax[0].yaxis.set_major_locator(plt_autotick()) ax[0].set_xlabel('sample value') ax[0].set_ylabel('frequency') ax[0].set_title('') ax[0].grid(axis='both') ax[1].set_xlabel('iteration') ax[1].set_ylabel('sample value') ax[1].grid(axis='both') ax[1].set_title('') fig = ax[0].figure fig.suptitle(c_name) fig.savefig(fig_dir + out_fname + '_stan_traceplot_' + c_name + '_arviz' + '.png') return None
20,349
47.684211
128
py
ngmm_tools
ngmm_tools-master/Analyses/Prediction/create_scen_dataframe.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sat Aug 20 17:26:17 2022 @author: glavrent """ #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #geographic libraries import pyproj import geopy.distance #user libraries sys.path.insert(0,'../Python_lib/ground_motions') from pylib_gmm_eas import BA18 ba18 = BA18() # Define Problem # --------------------------- #structural period freq = 5.0119 #earthquake scenario mag = 7.0 vs30 = 400 sof = 'SS' dip = 90 z_tor = 0 #color bar limits cbar_lim = [np.log(1e-8),np.log(.06)] #earthquake coordinates scen_eq_latlon = [34.2, -116.9] #utm zone utm_zone = '11S' #grid grid_X_dxdy = [10, 10] #scenario filename fname_scen_predict = '../../Data/Prediction/scen_predict.csv' # UTM projection # --------------------------- # projection system utmProj = pyproj.Proj("+proj=utm +zone="+utm_zone+", +ellps=WGS84 +datum=WGS84 +units=m +no_defs") #grid limits in UTM grid_X_win = np.array([[-140, 3500], [780, 4700]]) #create coordinate grid grid_x_edge = np.arange(grid_X_win[0,0],grid_X_win[1,0],grid_X_dxdy[0]) grid_y_edge = np.arange(grid_X_win[0,1],grid_X_win[1,1],grid_X_dxdy[0]) grid_x, grid_y = np.meshgrid(grid_x_edge, grid_y_edge) #create coordinate array with all grid nodes grid_X = np.vstack([grid_x.T.flatten(), grid_y.T.flatten()]).T #compute lat/lon coordinate array grid_latlon = np.fliplr(np.array([utmProj(g_x*1000, g_y*1000, inverse=True) for g_x, g_y in zip(grid_X[:,0], grid_X[:,1])])) n_gpt = len(grid_X) #earthquake UTM coordinates scen_eq_X = np.array(utmProj(scen_eq_latlon[1], scen_eq_latlon[0])) / 1000 #create earthquake and site ids eqid_array = np.full(n_gpt, -1) site_array = -1*(1+np.arange(n_gpt)) # Compute Ergodic Base Scaling # --------------------------- #compute distances scen_dist_array = np.linalg.norm(grid_X - scen_eq_X, axis=1) scen_dist_array = np.sqrt(scen_dist_array**2 + z_tor**2) #scenarios of interest scen_eas_nerg_scl = np.full(n_gpt, np.nan) scen_eas_nerg_aleat = np.full(n_gpt, np.nan) for k, d in enumerate(scen_dist_array): fnorm = 1 if sof == 'SS' else 0 #median and aleatory scen_eas_nerg_scl[k], _, scen_eas_nerg_aleat[k] = ba18.EasF(freq, mag, rrup=d, vs30=vs30, ztor=z_tor, fnorm=fnorm, flag_keep_b7 = False) # Summarize Scenario Dataframe # --------------------------- df_scen_prdct = pd.DataFrame({'eqid':eqid_array, 'ssn':site_array, 'eqLat':np.full(n_gpt,scen_eq_latlon[0]), 'eqLon':np.full(n_gpt,scen_eq_latlon[0]), 'staLat':grid_latlon[:,0], 'staLon':grid_latlon[:,1], 'eqX':np.full(n_gpt,scen_eq_X[0]), 'eqY':np.full(n_gpt,scen_eq_X[1]), 'eqZ':np.full(n_gpt,-z_tor), 'staX':grid_X[:,0], 'staY':grid_X[:,1], 'erg_base':scen_eas_nerg_scl}) #save prediction scenarios df_scen_prdct.to_csv(fname_scen_predict )
3,049
28.61165
140
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/synthetic_datasets/create_synthetic_ds1.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Jul 1 21:25:34 2021 @author: glavrent """ # load libraries import os import pathlib # arithmetic libraries import numpy as np # statistics libraries import pandas as pd # python interface to Stan for Bayesian inference # for installation check https://pystan.readthedocs.io/en/latest/ import pystan # set working directories os.chdir(os.getcwd()) # change directory to current directory # %% Define Input Data # ====================================== # USER SETS THE INPUT FLATFILE NAMES AND PATH # ++++++++++++++++++++++++++++++++++++++++ #input flatfile # fname_flatfile = 'CatalogNGAWest3CA' # fname_flatfile = 'CatalogNGAWest3CA_2013' # fname_flatfile = 'CatalogNGAWest3NCA' # fname_flatfile = 'CatalogNGAWest3SCA' fname_flatfile = 'CatalogNGAWest3CALite' dir_flatfile = '../../../Data/Validation/preprocessing/flatfiles/merged/' # ++++++++++++++++++++++++++++++++++++++++ # USER SETS THE INPUT FLATFILE NAMES AND PATH # ++++++++++++++++++++++++++++++++++++++++ fname_stan_model = 'create_synthetic_ds1.stan' # ++++++++++++++++++++++++++++++++++++++++ # USER SETS THE OUTPUT FILE PATH AND NAME # ++++++++++++++++++++++++++++++++++++++++ # output filename sufix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' # output directories dir_out = f'../../../Data/Validation/synthetic_datasets/ds1{synds_suffix}/' # ++++++++++++++++++++++++++++++++++++++++ # user defines hyper parameters # -------------------------------------- # number of synthetic data-sets n_ds = 5 # number of chains and seed number in stan model n_chains = 1 n_seed = 1 # define hyper-parameters # omega_0: standard deviation for constant offset # omega_1e: standard deviation for spatially varying earthquake constant # omega_1as: standard deviation for spatially vayring site constant # omega_1bs: standard deviation for independent site constant # ell_1e: correlation lenght for spatially varying earthquake constant # ell_1as: correlation lenght for spatially vayring site constant # phi_0: within-event standard deviation # tau_0: between-event standard deviation # USER NEEDS TO SPECIFY HYPERPARAMETERS # ++++++++++++++++++++++++++++++++++++++++ # # small correlation lengths # hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, # 'ell_1e':60, 'ell_1as':30, 'phi_0':0.4, 'tau_0':0.3 } # # #large correlation lengths # hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, # 'ell_1e':100, 'ell_1as':70, 'phi_0':0.4, 'tau_0':0.3 } # ++++++++++++++++++++++++++++++++++++++++ # %% Load Data # ====================================== #load flatfile fullname_flatfile = dir_flatfile + fname_flatfile + '.csv' df_flatfile = pd.read_csv(fullname_flatfile) # %% Processing # ====================================== # read earthquake and station data from the flatfile n_rec = len(df_flatfile) # read earthquake data # earthquake IDs (eqid), magnitudes (mag), and coordinates (eqX,eqY) # user may change these IDs based on the headers of the flatfile data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_index=True, return_inverse=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] #earthquake coordinates # create earthquake ids for all recordings eq_id = eq_inv + 1 n_eq = len(data_eq) # read station data # station IDs (ssn), Vs30, and coordinates (staX,staY) # user may change these IDs based on the headers of the flatfile data_stat_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique(df_flatfile[['ssn']].values, axis = 0, return_index=True, return_inverse=True) data_stat = data_stat_all[sta_idx,:] X_stat = data_stat[:,[2,3]] #station coordinates # create station ids for all recordings sta_id = sta_inv + 1 n_stat = len(data_stat) # %% Stan # ====================================== ## Stan Data # --------------------------- stan_data = {'N': n_rec, 'NEQ': n_eq, 'NSTAT': n_stat, 'X_e': X_eq, #earthquake coordinates 'X_s': X_stat, #station coordinates 'eq': eq_id, #earthquake index 'stat': sta_id, #station index 'mu_gmm': np.zeros(n_rec), #hyper-parameters of generated data-set 'omega_0': hyp['omega_0'], 'omega_1e': hyp['omega_1e'], 'omega_1as': hyp['omega_1as'], 'omega_1bs': hyp['omega_1bs'], 'ell_1e': hyp['ell_1e'], 'ell_1as': hyp['ell_1as'], #aleatory terms 'phi_0': hyp['phi_0'], 'tau_0': hyp['tau_0'] } ## Compile and Run Stan model # --------------------------- # compile model sm = pystan.StanModel(file=fname_stan_model) # generate samples fit = sm.sampling(data=stan_data, algorithm="Fixed_param", iter=n_ds, chains=n_chains, seed=n_seed) # keep valid datasets Y_nerg_med = fit['Y_nerg_med'] Y_aleat = fit['Y_aleat'] Y_tot = fit['Y_tot'] # %% Output # ====================================== if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # save generated data-sets for k, (Y_nm, Y_t) in enumerate(zip(Y_nerg_med, Y_tot)): #copy catalog info to synthetic data-set df_synthetic_data = df_flatfile.copy() #add residuals columns df_synthetic_data.loc[:,'nerg_gm'] = Y_nm df_synthetic_data.loc[:,'tot'] = Y_t #add columns with sampled coefficients df_synthetic_data.loc[:,'dc_0'] = fit['dc_0'][k] df_synthetic_data.loc[:,'dc_1e'] = fit['dc_1e'][k][eq_inv] df_synthetic_data.loc[:,'dc_1as'] = fit['dc_1as'][k][sta_inv] df_synthetic_data.loc[:,'dc_1bs'] = fit['dc_1bs'][k][sta_inv] #add columns aleatory terms df_synthetic_data.loc[:,'dW'] = fit['dW'][k] df_synthetic_data.loc[:,'dB'] = fit['dB'][k][eq_inv] #create data-frame with synthetic dataset fname_synthetic_data = dir_out + f'{fname_flatfile}_synthetic_data{synds_suffix}_Y{k+1}.csv' df_synthetic_data.to_csv(fname_synthetic_data, index=False)
6,252
34.528409
110
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/synthetic_datasets/create_synthetic_ds3.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Dec 26 15:47:17 2021 @author: glavrent """ # load libraries import os import pathlib # arithmetic libraries import numpy as np # statistics libraries import pandas as pd # python interface to Stan for Bayesian inference # for installation check https://pystan.readthedocs.io/en/latest/ import pystan # %% set working directories os.chdir(os.getcwd()) # change directory to current directory # %% Define Input Data # ====================================== # USER SETS THE INPUT FLATFILE NAMES AND PATH # ++++++++++++++++++++++++++++++++++++++++ #input flatfile # fname_flatfile = 'CatalogNGAWest3CA' # fname_flatfile = 'CatalogNGAWest3CA_2013' # fname_flatfile = 'CatalogNGAWest3NCA' # fname_flatfile = 'CatalogNGAWest3SCA' fname_flatfile = 'CatalogNGAWest3CALite' dir_flatfile = '../../../Data/Validation/preprocessing/flatfiles/merged/' # cell data # fname_cellinfo = 'CatalogNGAWest3CA_cellinfo.csv' # fname_celldist = 'CatalogNGAWest3CA_distancematrix.csv' # fname_cellinfo = 'CatalogNGAWest3CA_2013_cellinfo.csv' # fname_celldist = 'CatalogNGAWest3CA_2013_distancematrix.csv' fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo.csv' fname_celldist = 'CatalogNGAWest3CALite_distancematrix.csv' fname_celldist_sp = 'CatalogNGAWest3CALite_distancematrix_sparce.csv' dir_celldata = '../../../Data/Validation/preprocessing/cell_distances/' # ++++++++++++++++++++++++++++++++++++++++ # USER SETS THE INPUT FLATFILE NAMES AND PATH # ++++++++++++++++++++++++++++++++++++++++ fname_stan_model = 'create_synthetic_ds3.stan' # ++++++++++++++++++++++++++++++++++++++++ # USER SETS THE OUTPUT FILE PATH AND NAME # ++++++++++++++++++++++++++++++++++++++++ # output filename sufix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' # output directories dir_out = f'../../../Data/Validation/synthetic_datasets/ds3{synds_suffix}/' # ++++++++++++++++++++++++++++++++++++++++ # number of synthetic data-sets n_dataset = 5 n_attempts = 500 # number of chains and seed number in stan model n_chains = 1 n_seed = 1 # define hyper-parameters # omega_0: standard deviation for constant offset # omega_1e: standard deviation for spatially varying earthquake constant # omega_1as: standard deviation for spatially varying site constant # omega_1bs: standard deviation for independent site constant # ell_1e: correlation length for spatially varying earthquake constant # ell_1as: correlation length for spatially varying site constant # c_2_erg: ergodic geometrical-spreading coefficient # omega_2: standard deviation for shift in average geometrical-spreading # omega_2p: standard deviation for spatially varying geometrical-spreading coefficient # ell_2p: correlation length for spatially varying geometrical-spreading coefficient # c_3_erg: ergodic Vs30 scaling coefficient # omega_3: standard deviation for shift in average Vs30 scaling # omega_3s: standard deviation for spatially varying Vs30 scaling # ell_3s: correlation length for spatially varying Vs30 scaling # c_cap_erg: erogodic cell-specific anelastic attenuation # omega_cap_mu: standard deviation for constant offset of cell-specific anelastic attenuation # omega_ca1p: standard deviation for spatially varying component of cell-specific anelastic attenuation # omega_ca2p: standard deviation for spatially independent component of cell-specific anelastic attenuation # ell_ca1p: correlation length for spatially varying component of cell-specific anelastic attenuation # phi_0: within-event standard deviation # tau_0: between-event standard deviation # USER NEEDS TO SPECIFY HYPERPARAMETERS # ++++++++++++++++++++++++++++++++++++++++ # # small correlation lengths # hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, # 'ell_1e':60, 'ell_1as':30, # 'c_2_erg': -2.0, # 'omega_2': 0.2, # 'omega_2p': 0.15, 'ell_2p': 80, # 'c_3_erg':-0.6, # 'omega_3': 0.15, # 'omega_3s': 0.15, 'ell_3s': 130, # 'c_cap_erg': -0.011, # 'omega_cap_mu': 0.005, 'omega_ca1p':0.004, 'omega_ca2p':0.002, # 'ell_ca1p': 75, # 'phi_0':0.3, 'tau_0':0.25 } # # large correlation lengths # hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, # 'ell_1e':100, 'ell_1as':70, # 'c_2_erg': -2.0, # 'omega_2': 0.2, # 'omega_2p': 0.15, 'ell_2p': 140, # 'c_3_erg':-0.6, # 'omega_3': 0.15, # 'omega_3s': 0.15, 'ell_3s': 180, # 'c_cap_erg': -0.02, # 'omega_cap_mu': 0.008, 'omega_ca1p':0.005, 'omega_ca2p':0.003, # 'ell_ca1p': 120, # 'phi_0':0.3, 'tau_0':0.25} # ++++++++++++++++++++++++++++++++++++++++ #psuedo depth term for mag saturation h_M = 4 # %% Load Data # ====================================== # load flatfile fullname_flatfile = dir_flatfile + fname_flatfile + '.csv' df_flatfile = pd.read_csv(fullname_flatfile) # load celldata df_cell_dist = pd.read_csv(dir_celldata + fname_celldist, index_col=0) df_cell_dist_sp = pd.read_csv(dir_celldata + fname_celldist_sp) df_cell_info = pd.read_csv(dir_celldata + fname_cellinfo) # %% Processing # ====================================== # read earthquake and station data from the flatfile n_rec = len(df_flatfile) # read earthquake data # earthquake IDs (rqid), magnitudes (mag), and coordinates (eqX,eqY) # user may change these IDs based on the headers of the flatfile data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_index=True, return_inverse=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] # earthquake coordinates # create earthquake ids for all recordings eq_id = eq_inv + 1 n_eq = len(data_eq) # read station data # station IDs (ssn), Vs30, and coordinates (staX,staY) # user may change these IDs based on the headers of the flatfile data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique(df_flatfile[['ssn']].values, axis = 0, return_index=True, return_inverse=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] # station coordinates # create station ids for all recordings sta_id = sta_inv + 1 n_sta = len(data_sta) # geometrical spreading covariate x_2 = np.log(np.sqrt(df_flatfile.Rrup.values**2 + h_M**2)) #vs30 covariate x_3 = np.log(np.minimum(data_sta[:,1], 1000)/1000) assert(~np.isnan(x_3).all()),'Error. Invalid Vs30 values' # read cell data n_cell = len(df_cell_info) df_cell_dist = df_cell_dist.reindex(df_flatfile.rsn) #cell distance matrix for records in the synthetic data-set # cell names cells_names = df_cell_info.cellname.values cells_id = df_cell_info.cellid.values # cell distance matrix cell_dmatrix = df_cell_dist.loc[:,cells_names].values # cell coordinates X_cell = df_cell_info[['mptX','mptY']].values # valid cells i_val_cells = cell_dmatrix.sum(axis=0) > 0 # %% Stan # ====================================== ## Stan Data # --------------------------- stan_data = {'N': n_rec, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell, 'eq': eq_id, #earthquake index 'stat': sta_id, #station index 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cell, #cell coordinates 'RC': cell_dmatrix, #cell distances 'mu_gmm': np.zeros(n_rec), #covariates 'x_2': x_2, #geometrical spreading 'x_3': x_3, #Vs30 scaling #hyper-parameters of generated data-set 'omega_0': hyp['omega_0'], 'omega_1e': hyp['omega_1e'], 'omega_1as': hyp['omega_1as'], 'omega_1bs': hyp['omega_1bs'], 'ell_1e': hyp['ell_1e'], 'ell_1as': hyp['ell_1as'], 'c_2_erg': hyp['c_2_erg'], 'omega_2': hyp['omega_2'], 'omega_2p': hyp['omega_2p'], 'ell_2p': hyp['ell_2p'], 'c_3_erg': hyp['c_3_erg'], 'omega_3': hyp['omega_3'], 'omega_3s': hyp['omega_3s'], 'ell_3s': hyp['ell_3s'], #anelastic attenuation 'c_cap_erg': hyp['c_cap_erg'], 'omega_cap_mu': hyp['omega_cap_mu'], 'omega_ca1p': hyp['omega_ca1p'], 'omega_ca2p': hyp['omega_ca2p'], 'ell_ca1p': hyp['ell_ca1p'], #aleatory terms 'phi_0': hyp['phi_0'], 'tau_0': hyp['tau_0'] } ## Compile and Run Stan model # --------------------------- # compile model sm = pystan.StanModel(file=fname_stan_model) # generate samples fit = sm.sampling(data=stan_data, algorithm="Fixed_param", iter=n_attempts, chains=n_chains, seed=n_seed) # select only data-sets with negative anelastic attenuation coefficients valid_dataset = np.array( n_attempts * [False] ) for k, (c_2p, c_cap) in enumerate(zip(fit['c_2p'], fit['c_cap'])): valid_dataset[k] = np.all(c_2p <= 0 ) & np.all(c_cap <= 0 ) valid_dataset = np.where(valid_dataset)[0] #valid data-set ids valid_dataset = valid_dataset[:min(n_dataset,len(valid_dataset))] # keep valid datasets Y_nerg_med = fit['Y_nerg_med'][valid_dataset] Y_var_coeff = fit['Y_var_ceoff'][valid_dataset] Y_inattent = fit['Y_inattent'][valid_dataset] Y_aleat = fit['Y_aleat'][valid_dataset] Y_tot = fit['Y_tot'][valid_dataset] c_cap = fit['c_cap'][valid_dataset] # %% Output # ====================================== if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # save generated data-sets for k, (k_vds, Y_nm, Y_vc, Y_iatt, Y_t) in enumerate(zip(valid_dataset, Y_nerg_med, Y_var_coeff, Y_inattent, Y_tot)): #copy catalog info to synthetic data-set df_synthetic_data = df_flatfile.copy() #add covariates df_synthetic_data.loc[:,'x_2'] = x_2 df_synthetic_data.loc[:,'x_3'] = x_3[sta_inv] #add residuals columns df_synthetic_data.loc[:,'nerg_gm'] = Y_nm df_synthetic_data.loc[:,'vcoeff'] = Y_vc df_synthetic_data.loc[:,'inatten'] = Y_iatt df_synthetic_data.loc[:,'tot'] = Y_t #add columns with sampled coefficients df_synthetic_data.loc[:,'dc_0'] = fit['dc_0'][k_vds] df_synthetic_data.loc[:,'dc_1e'] = fit['dc_1e'][k_vds][eq_inv] df_synthetic_data.loc[:,'dc_1as'] = fit['dc_1as'][k_vds][sta_inv] df_synthetic_data.loc[:,'dc_1bs'] = fit['dc_1bs'][k_vds][sta_inv] df_synthetic_data.loc[:,'c_2'] = fit['c_2_mu'][k_vds] df_synthetic_data.loc[:,'c_2p'] = fit['c_2p'][k_vds][eq_inv] df_synthetic_data.loc[:,'c_3'] = fit['c_3_mu'][k_vds] df_synthetic_data.loc[:,'c_3s'] = fit['c_3s'][k_vds][sta_inv] #add columns aleatory terms df_synthetic_data.loc[:,'dW'] = fit['dW'][k_vds] df_synthetic_data.loc[:,'dB'] = fit['dB'][k_vds][eq_inv] #create data-frame with synthetic dataset fname_synthetic_data = dir_out + f'{fname_flatfile}_synthetic_data{synds_suffix}_Y{k+1}.csv' df_synthetic_data.to_csv(fname_synthetic_data, index=False) # save coeffiicients for k, (k_vds, c_ca) in enumerate(zip(valid_dataset, c_cap)): #create synthetic cell dataset df_synthetic_cell = df_cell_info.copy() #cell specific anelastic attenuation df_synthetic_cell.loc[:,'c_cap_mu'] = fit['c_cap_mu'][k_vds] df_synthetic_cell.loc[:,'c_cap'] = c_ca #create data-frame with cell specific dataset fname_synthetic_atten = dir_out + f'{fname_flatfile}_synthetic_atten{synds_suffix}_Y{k+1}.csv' df_synthetic_cell.to_csv(fname_synthetic_atten, index=False) # save cell data fname_cell_info = dir_out + f'{fname_flatfile}_cellinfo.csv' fname_cell_dist = dir_out + f'{fname_flatfile}_distancematrix.csv' fname_cell_dist_sp = dir_out + f'{fname_flatfile}_distancematrix_sparse.csv' df_cell_info.to_csv(fname_cell_info, index=False) df_cell_dist.to_csv(fname_cell_dist) df_cell_dist_sp.to_csv(fname_cell_dist_sp, index=False)
12,361
40.206667
117
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/synthetic_datasets/create_synthetic_ds2.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Jul 1 21:25:34 2021 @author: glavrent """ # load libraries import os import pathlib # arithmetic libraries import numpy as np # statistics libraries import pandas as pd # python interface to Stan for Bayesian inference # for installation check https://pystan.readthedocs.io/en/latest/ import pystan # %% set working directories os.chdir(os.getcwd()) # change directory to current directory # %% Define Input Data # ====================================== # USER SETS THE INPUT FLATFILE NAMES AND PATH # ++++++++++++++++++++++++++++++++++++++++ #input flatfile # fname_flatfile = 'CatalogNGAWest3CA' # fname_flatfile = 'CatalogNGAWest3CA_2013' # fname_flatfile = 'CatalogNGAWest3NCA' # fname_flatfile = 'CatalogNGAWest3SCA' fname_flatfile = 'CatalogNGAWest3CALite' dir_flatfile = '../../../Data/Validation/preprocessing/flatfiles/merged/' # cell data # fname_cellinfo = 'CatalogNGAWest3CA_cellinfo.csv' # fname_celldist = 'CatalogNGAWest3CA_distancematrix.csv' # fname_cellinfo = 'CatalogNGAWest3CA_2013_cellinfo.csv' # fname_celldist = 'CatalogNGAWest3CA_2013_distancematrix.csv' fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo.csv' fname_celldist = 'CatalogNGAWest3CALite_distancematrix.csv' fname_celldist_sp = 'CatalogNGAWest3CALite_distancematrix_sparce.csv' dir_celldata = '../../../Data/Validation/preprocessing/cell_distances/' # ++++++++++++++++++++++++++++++++++++++++ # USER SETS THE INPUT FLATFILE NAMES AND PATH # ++++++++++++++++++++++++++++++++++++++++ fname_stan_model = 'create_synthetic_ds2.stan' # ++++++++++++++++++++++++++++++++++++++++ # USER SETS THE OUTPUT FILE PATH AND NAME # ++++++++++++++++++++++++++++++++++++++++ # output filename sufix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' # output directories dir_out = f'../../../Data/Validation/synthetic_datasets/ds2{synds_suffix}/' # ++++++++++++++++++++++++++++++++++++++++ # number of synthetic data-sets n_dataset = 5 n_attempts = 500 # number of chains and seed number in stan model n_chains = 1 n_seed = 1 # define hyper-parameters # omega_0: standard deviation for constant offset # omega_1e: standard deviation for spatially varying earthquake constant # omega_1as: standard deviation for spatially varying site constant # omega_1bs: standard deviation for independent site constant # ell_1e: correlation length for spatially varying earthquake constant # ell_1as: correlation length for spatially varying site constant # c_cap_erg: erogodic cell-specific anelastic attenuation # omega_cap_mu: standard deviation for constant offset of cell-specific anelastic attenuation # omega_ca1p: standard deviation for spatially varying component of cell-specific anelastic attenuation # omega_ca2p: standard deviation for spatially independent component of cell-specific anelastic attenuation # ell_ca1p: correlation length for spatially varying component of cell-specific anelastic attenuation # phi_0: within-event standard deviation # tau_0: between-event standard deviation # USER NEEDS TO SPECIFY HYPERPARAMETERS # ++++++++++++++++++++++++++++++++++++++++ # # small correlation lengths # hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, # 'ell_1e':60, 'ell_1as':30, # 'c_cap_erg': -0.011, # 'omega_cap_mu': 0.005, 'omega_ca1p':0.004, 'omega_ca2p':0.002, # 'ell_ca1p': 75, # 'phi_0':0.4, 'tau_0':0.3 } # # # large correlation lengths # hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, # 'ell_1e':100, 'ell_1as':70, # 'c_cap_erg': -0.02, # 'omega_cap_mu': 0.008, 'omega_ca1p':0.005, 'omega_ca2p':0.003, # 'ell_ca1p': 120, # 'phi_0':0.4, 'tau_0':0.3} # ++++++++++++++++++++++++++++++++++++++++ # %% Load Data # ====================================== # load flatfile fullname_flatfile = dir_flatfile + fname_flatfile + '.csv' df_flatfile = pd.read_csv(fullname_flatfile) # load celldata df_cell_dist = pd.read_csv(dir_celldata + fname_celldist, index_col=0) df_cell_dist_sp = pd.read_csv(dir_celldata + fname_celldist_sp) df_cell_info = pd.read_csv(dir_celldata + fname_cellinfo) # %% Processing # ====================================== # read earthquake and station data from the flatfile n_rec = len(df_flatfile) # read earthquake data # earthquake IDs (rqid), magnitudes (mag), and coordinates (eqX,eqY) # user may change these IDs based on the headers of the flatfile data_eq_all = df_flatfile[['eqid','mag','eqX', 'eqY']].values _, eq_idx, eq_inv = np.unique(df_flatfile[['eqid']], axis=0, return_index=True, return_inverse=True) data_eq = data_eq_all[eq_idx,:] X_eq = data_eq[:,[2,3]] # earthquake coordinates # create earthquake ids for all recordings eq_id = eq_inv + 1 n_eq = len(data_eq) # read station data # station IDs (ssn), Vs30, and coordinates (staX,staY) # user may change these IDs based on the headers of the flatfile data_sta_all = df_flatfile[['ssn','Vs30','staX','staY']].values _, sta_idx, sta_inv = np.unique(df_flatfile[['ssn']].values, axis = 0, return_index=True, return_inverse=True) data_sta = data_sta_all[sta_idx,:] X_sta = data_sta[:,[2,3]] # station coordinates # create station ids for all recordings sta_id = sta_inv + 1 n_sta = len(data_sta) # read cell data n_cell = len(df_cell_info) df_cell_dist = df_cell_dist.reindex(df_flatfile.rsn) #cell distance matrix for records in the synthetic data-set # cell names cells_names = df_cell_info.cellname.values cells_id = df_cell_info.cellid.values # cell distance matrix cell_dmatrix = df_cell_dist.loc[:,cells_names].values # cell coordinates X_cell = df_cell_info[['mptX','mptY']].values # valid cells i_val_cells = cell_dmatrix.sum(axis=0) > 0 # %% Stan # ====================================== ## Stan Data # --------------------------- stan_data = {'N': n_rec, 'NEQ': n_eq, 'NSTAT': n_sta, 'NCELL': n_cell, 'eq': eq_id, #earthquake index 'stat': sta_id, #station index 'X_e': X_eq, #earthquake coordinates 'X_s': X_sta, #station coordinates 'X_c': X_cell, #cell coordinates 'RC': cell_dmatrix, #cell distances 'mu_gmm': np.zeros(n_rec), #hyper-parameters of generated data-set 'omega_0': hyp['omega_0'], 'omega_1e': hyp['omega_1e'], 'omega_1as': hyp['omega_1as'], 'omega_1bs': hyp['omega_1bs'], 'ell_1e': hyp['ell_1e'], 'ell_1as': hyp['ell_1as'], #anelastic attenuation 'c_cap_erg': hyp['c_cap_erg'], 'omega_cap_mu': hyp['omega_cap_mu'], 'omega_ca1p': hyp['omega_ca1p'], 'omega_ca2p': hyp['omega_ca2p'], 'ell_ca1p': hyp['ell_ca1p'], #aleatory terms 'phi_0': hyp['phi_0'], 'tau_0': hyp['tau_0'] } ## Compile and Run Stan model # --------------------------- # compile model sm = pystan.StanModel(file=fname_stan_model) # generate samples fit = sm.sampling(data=stan_data, algorithm="Fixed_param", iter=n_attempts, chains=n_chains, seed=n_seed) # select only data-sets with negative anelastic attenuation coefficients valid_dataset = np.array( n_attempts * [False] ) for k, c_cap in enumerate(fit['c_cap']): valid_dataset[k] = np.all(c_cap <= 0 ) valid_dataset = np.where(valid_dataset)[0] #valid data-set ids valid_dataset = valid_dataset[:min(n_dataset,len(valid_dataset))] # keep valid datasets Y_nerg_med = fit['Y_nerg_med'][valid_dataset] Y_var_coeff = fit['Y_var_ceoff'][valid_dataset] Y_inattent = fit['Y_inattent'][valid_dataset] Y_aleat = fit['Y_aleat'][valid_dataset] Y_tot = fit['Y_tot'][valid_dataset] c_cap = fit['c_cap'][valid_dataset] # %% Output # ====================================== if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # save generated data-sets for k, (k_vds, Y_nm, Y_vc, Y_iatt, Y_t) in enumerate(zip(valid_dataset, Y_nerg_med, Y_var_coeff, Y_inattent, Y_tot)): #copy catalog info to synthetic data-set df_synthetic_data = df_flatfile.copy() #add residuals columns df_synthetic_data.loc[:,'nerg_gm'] = Y_nm df_synthetic_data.loc[:,'vcoeff'] = Y_vc df_synthetic_data.loc[:,'inatten'] = Y_iatt df_synthetic_data.loc[:,'tot'] = Y_t #add columns with sampled coefficients df_synthetic_data.loc[:,'dc_0'] = fit['dc_0'][k_vds] df_synthetic_data.loc[:,'dc_1e'] = fit['dc_1e'][k_vds][eq_inv] df_synthetic_data.loc[:,'dc_1as'] = fit['dc_1as'][k_vds][sta_inv] df_synthetic_data.loc[:,'dc_1bs'] = fit['dc_1bs'][k_vds][sta_inv] #add columns aleatory terms df_synthetic_data.loc[:,'dW'] = fit['dW'][k_vds] df_synthetic_data.loc[:,'dB'] = fit['dB'][k_vds][eq_inv] #create data-frame with synthetic dataset fname_synthetic_data = dir_out + f'{fname_flatfile}_synthetic_data{synds_suffix}_Y{k+1}.csv' df_synthetic_data.to_csv(fname_synthetic_data, index=False) # save coeffiicients for k, (k_vds, c_ca) in enumerate(zip(valid_dataset, c_cap)): #create synthetic cell dataset df_synthetic_cell = df_cell_info.copy() #cell specific anelastic attenuation df_synthetic_cell.loc[:,'c_cap_mu'] = fit['c_cap_mu'][k_vds] df_synthetic_cell.loc[:,'c_cap'] = c_ca #create data-frame with cell specific dataset fname_synthetic_atten = dir_out + f'{fname_flatfile}_synthetic_atten{synds_suffix}_Y{k+1}.csv' df_synthetic_cell.to_csv(fname_synthetic_atten, index=False) # save cell data fname_cell_info = dir_out + f'{fname_flatfile}_cellinfo.csv' fname_cell_dist = dir_out + f'{fname_flatfile}_distancematrix.csv' fname_cell_dist_sp = dir_out + f'{fname_flatfile}_distancematrix_sparse.csv' df_cell_info.to_csv(fname_cell_info, index=False) df_cell_dist.to_csv(fname_cell_dist) df_cell_dist_sp.to_csv(fname_cell_dist_sp, index=False)
10,211
38.890625
117
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/main_pystan_model1_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model1_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds1' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info # ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #stan model # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds1/' #output sub-directory #pystan2 # out_dir_sub = 'PYSTAN_NGAWest2CANorth' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_chol' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_chol_eff2' #pystan3 # out_dir_sub = 'PYSTAN3_NGAWest2CANorth' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_chol' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_chol_eff' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_chol_eff2' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, sm_fname, out_fname, out_dir, res_name, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,687
30.793103
90
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/comparison_stan_model1.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Aug 12 20:52:09 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load packages import os import sys import pathlib import glob import re #regular expression package import pickle #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick #user functions sys.path.insert(0,'../../../Python_lib/regression/') from pylib_stats import CalcRMS from pylib_stats import CalcLKDivergece # Define variables # --------------------------- # USER SETS DIRECTORIES AND FILE INFO OF SYNTHETIC DS AND REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #processed dataset # name_dataset = 'NGAWest2CANorth' name_dataset = 'NGAWest2CA' # name_dataset = 'NGAWest3CA' #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #package # 1: Pystan v2 # 2: Pystan v3 # 3: stancmd pkg_id = 3 #approximation type # 1: multivariate normal # 2: cholesky # 3: cholesky efficient # 4: cholesky efficient v2 aprox_id = 3 #directories (synthetic dataset) if corr_id == 1: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds1_small_corr_len' elif corr_id == 2: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds1_large_corr_len' #directories (regression results) if pkg_id == 1: dir_results = f'../../../../Data/Verification/regression/ds1/PYSTAN_%s'%name_dataset elif pkg_id == 2: dir_results = f'../../../../Data/Verification/regression/ds1/PYSTAN3_%s'%name_dataset elif pkg_id == 3: dir_results = f'../../../../Data/Verification/regression/ds1/CMDSTAN_%s'%name_dataset #prefix for synthetic data and results prfx_syndata = 'CatalogNGAWest3CALite_synthetic' # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #regression results filename prefix prfx_results = f'%s_syndata'%name_dataset #output filename sufix (synthetic dataset) if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #output filename sufix (regression results) if aprox_id == 1: synds_suffix_stan = synds_suffix elif aprox_id == 2: synds_suffix_stan = '_chol' + synds_suffix elif aprox_id == 3: synds_suffix_stan = '_chol_eff' + synds_suffix elif aprox_id == 4: synds_suffix_stan = '_chol_eff2' + synds_suffix # dataset info ds_id = np.arange(1,6) # USER NEEDS TO SPECIFY HYPERPARAMETERS OF SYNTHETIC DATASET # ++++++++++++++++++++++++++++++++++++++++ # hyper-parameters if corr_id == 1: # small correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, 'ell_1e':60, 'ell_1as':30, 'phi_0':0.4, 'tau_0':0.3 } elif corr_id == 2: #large correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, 'ell_1e':100, 'ell_1as':70, 'phi_0':0.4, 'tau_0':0.3 } # ++++++++++++++++++++++++++++++++++++++++ #ploting options flag_report = True # Compare coefficients # --------------------------- #initialize misfit metrics dataframe df_misfit = pd.DataFrame(index=['Y%i'%d_id for d_id in ds_id]) #iterate over different datasets for d_id in ds_id: # Load Data #--- --- --- --- --- #file names #synthetic data fname_sdata_gmotion = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'data', synds_suffix, d_id) + '.csv' #regression results fname_reg_gmotion = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'residuals') + '.csv' fname_reg_coeff = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'coefficients') + '.csv' #load synthetic results df_sdata_gmotion = pd.read_csv(fname_sdata_gmotion).set_index('rsn') #load regression results df_reg_gmotion = pd.read_csv(fname_reg_gmotion, index_col=0) df_reg_coeff = pd.read_csv(fname_reg_coeff, index_col=0) # Processing #--- --- --- --- --- #keep only common records from synthetic dataset df_sdata_gmotion = df_sdata_gmotion.reindex(df_reg_gmotion.index) #find unique earthqakes and stations eq_id, eq_idx, eq_nrec = np.unique(df_sdata_gmotion.eqid, return_index=True, return_counts=True) sta_id, sta_idx, sta_nrec = np.unique(df_sdata_gmotion.ssn, return_index=True, return_counts=True) # Compute Root Mean Square Error #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_rms'] = CalcRMS(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_rms'] = CalcRMS(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_rms'] = CalcRMS(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_rms'] = CalcRMS(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) # Compute Divergence #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_KL'] = CalcLKDivergece(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) # Output #--- --- --- --- --- #figure directory dir_fig = '%s%s/Y%i/figures_cmp/'%(dir_results,synds_suffix_stan,d_id) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #compare ground motion predictions #... ... ... ... ... ... #figure title fname_fig = 'Y%i_tot_res_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #median ax.scatter(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title('Comparison total residuals, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=35) ax.set_ylabel('Estimated', fontsize=35) ax.grid(which='both') ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2,2]) ax.set_ylim([-2,2]) fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1e #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1e_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,e}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-.4,.4]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['dc_1e_sig'].values[eq_idx], df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,e}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,e}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1as #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1as_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1as_sig'].values[sta_idx], df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1bs #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1bs_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1bs_sig'].values[sta_idx], df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare Misfit Metrics # --------------------------- #summary directory dir_sum = '%s%s/summary/'%(dir_results,synds_suffix_stan) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #figure directory dir_fig = '%s/figures/'%(dir_sum) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save df_misfit.to_csv(dir_sum + 'misfit_summary.csv') #RMS misfit fname_fig = 'misfit_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,e}$') ax.plot(ds_id, df_misfit.dc_1as_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,s}$') ax.plot(ds_id, df_misfit.dc_1bs_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,s}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('RSME', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #KL divergence fname_fig = 'KLdiv_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,e}$') ax.plot(ds_id, df_misfit.dc_1as_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,s}$') ax.plot(ds_id, df_misfit.dc_1bs_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,s}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('KL divergence', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare hyper-paramters # --------------------------- #iterate over different datasets df_reg_hyp = list() df_reg_hyp_post = list() for d_id in ds_id: # Load Data #--- --- --- --- --- #regression hyperparamters results fname_reg_hyp = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperparameters') + '.csv' fname_reg_hyp_post = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperposterior') + '.csv' #load regression results df_reg_hyp.append( pd.read_csv(fname_reg_hyp, index_col=0) ) df_reg_hyp_post.append( pd.read_csv(fname_reg_hyp_post, index_col=0) ) # Omega_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 40 ymax_mean = 40 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mean, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1,e}$', fontsize=30) ax.set_xlabel('$\omega_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.25]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 30 ymax_mean = 30 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1a,s}$', fontsize=30) ax.set_xlabel('$\omega_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1bs #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1bs' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1b,s}$', fontsize=30) ax.set_xlabel('$\omega_{1b,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 0.02 ymax_mean = 0.02 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1,e}$', fontsize=30) ax.set_xlabel('$\ell_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1a,s}$', fontsize=30) ax.set_xlabel('$\ell_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Tau_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'tau_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\tau_{0}$', fontsize=30) ax.set_xlabel(r'$\tau_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Phi_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'phi_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 100 ymax_mean = 100 #plot posterior dist ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\phi_{0}$', fontsize=30) ax.set_xlabel('$\phi_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.6]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # # Delta c_0 # #--- --- --- --- --- # #hyper-paramter name # name_hyp = 'dc_0' # #figure title # fname_fig = 'post_dist_' + name_hyp # #create figure # fig, ax = plt.subplots(figsize = (10,10)) # for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): # #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mean = 15 # #plot posterior dist # pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') # #plot true value # ymax_hyp = ymax_mean # # ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') # #edit figure # ax.set_title(r'Comparison $\delta c_{0}$', fontsize=30) # ax.set_xlabel('$\delta c_{0}$', fontsize=25) # ax.set_ylabel('probability density function ', fontsize=25) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=22) # ax.tick_params(axis='y', labelsize=22) # #plot limits # ax.set_xlim([-1,1]) # ax.set_ylim([0,ymax_hyp]) # #save figure # fig.tight_layout() # # fig.savefig( dir_fig + fname_fig + '.png' )
29,388
37.977454
153
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/comparison_inla_model1_time.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 15 22:38:50 2022 @author: glavrent """ # Working directory and Packages # --------------------------- #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick # Define variables # --------------------------- #mesh info mesh_info = ['coarse', 'medium', 'fine'] #dataset name dataset_name = ['NGAWest2CANorth', 'NGAWest2CA', 'NGAWest3CA'] #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #correlation name if corr_id == 1: synds_name = 'small corr len' synds_suffix = '_small_corr_len' elif corr_id == 2: synds_name = 'large corr len' synds_suffix = '_large_corr_len' #directories regressions dir_reg = '../../../../Data/Verification/regression/ds1/' #directory output dir_out = '../../../../Data/Verification/regression/ds1/comparisons/' # Load Data # --------------------------- #initialize dataframe df_runinfo_all = {}; #iterate over different analyses for j1, m_i in enumerate(mesh_info): for j2, d_n in enumerate(dataset_name): key_runinfo = '%s_%s'%(m_i, d_n) fname_runinfo = '%s/INLA_%s_%s%s/run_info.csv'%(dir_reg, d_n, m_i, synds_suffix) #store calc time df_runinfo_all[key_runinfo] = pd.read_csv(fname_runinfo) # Comparison Figures # --------------------------- pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) #line style (iterate with mesh info) line_style = [':','--','-'] #color map (iterate with dataset) c_map = plt.get_cmap('Dark2') #run time figure fig_fname = 'run_time_inla' #create figure axes fig, ax = plt.subplots(figsize = (20,10)) #iterate over different analyses for j2, d_n in enumerate(dataset_name): for j1, (m_i, l_s) in enumerate(zip(mesh_info, line_style)): key_runinfo = '%s_%s'%(m_i, d_n) # ds_id = df_runinfo_all[key_runinfo].ds_id ds_name = ['Y%i'%d_i for d_i in ds_id] # run_time = df_runinfo_all[key_runinfo].run_time ax.plot(ds_id, run_time, linestyle=l_s, marker='o', linewidth=2, markersize=10, color=c_map(j2), label='%s - %s'%(d_n, m_i)) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=30) ax.set_ylabel('Run Time (min)', fontsize=30) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=ds_name) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) #legend ax.legend(loc='center left', bbox_to_anchor=(1, 0.5), fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_out + fig_fname + '.png' )
2,861
25.747664
132
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/comparison_stan_inla_model1_misfit.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Fri Jun 10 15:40:29 2022 @author: glavrent """ # Working directory and Packages # --------------------------- #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import FormatStrFormatter #user functions def PlotRSMCmp(df_list, names_list, c_name, width, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) # x_offset = #plot rms value for nm, df_l in zip(names_list, df_list): ax.bar(np.arange(df_l)-x_offset, df_sum_reg_stan.nerg_tot_rms.values, width=width, label=nm) #figure properties ax.set_ylim([0, 0.2]) ax.set_xticks(labels=df_l.ds_name) ax.set_xlabel('Dataset', fontsize=35) ax.set_ylabel('RMSE', fontsize=35) ax.grid(which='both') ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) ax.yaxis.set_major_formatter(FormatStrFormatter('%.2f')) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) # Define variables # --------------------------- # COMPARISONS # Different Mesh sizes # --- --- --- --- --- cmp_name = 'STAN_INLA_medium' reg_title = [f'STAN - NGAW2 CA, North', f'STAN - NGAW2 CA', f'STAN - NGAW3* CA', f'INLA - NGAW2 CA, North', f'INLA - NGAW2 CA', f'INLA - NGAW3* CA' ] reg_fname = ['CMDSTAN_NGAWest2CANorth_chol_eff_small_corr_len', 'CMDSTAN_NGAWest2CA_chol_eff_small_corr_len', 'CMDSTAN_NGAWest3CA_chol_eff_small_corr_len', 'INLA_NGAWest2CANorth_medium_small_corr_len', 'INLA_NGAWest2CA_medium_small_corr_len', 'INLA_NGAWest3CA_medium_small_corr_len'] ds_name = ['NGAW2\nCA, North','NGAW2\nCA', 'NGAW3\nCA', 'NGAW2\nCA, North','NGAW2\nCA', 'NGAW3*\nCA'] ds_id = np.array([1,2,3,1,2,3]) sftwr_name = 3*['STAN'] + 3*['INLA'] sftwr_id = np.array(3*[1]+3*[2]) #directories regressions reg_dir = [f'../../../../Data/Verification/regression/ds1/%s/'%r_f for r_f in reg_fname] #directory output dir_out = '../../../../Data/Verification/regression/ds1/comparisons/' # Load Data # --------------------------- #intialize main regression summary dataframe df_sum_reg = pd.DataFrame({'ds_id':ds_id, 'ds_name':ds_name, 'sftwr_id':sftwr_id, 'sftwr_name':sftwr_name}) #initialize misfit dataframe df_sum_misfit_all = {}; #read misfit info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename misfit info fname_sum = r_d + 'summary/misfit_summary.csv' #read KL score for coefficients df_sum_misfit_all[r_t] = pd.read_csv(fname_sum, index_col=0) #summarize regression rms df_sum_reg.loc[k,'nerg_tot_rms'] = df_sum_misfit_all[r_t].nerg_tot_rms.mean() #initialize run time dataframe df_runinfo_all = {}; #read run time info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename run time fname_runinfo = r_d + '/run_info.csv' #store calc time df_runinfo_all[r_t] = pd.read_csv(fname_runinfo) #summarize regression rms df_sum_reg.loc[k,'run_time'] = df_runinfo_all[r_t].run_time.mean() #print mean run time print(f'%s: %.1f min'%( r_t, df_runinfo_all[r_t].run_time.mean() )) #separate STNA and INLA runs df_sum_reg_stan = df_sum_reg.loc[df_sum_reg.sftwr_id==1,:] df_sum_reg_inla = df_sum_reg.loc[df_sum_reg.sftwr_id==2,:] # Comparison Figures # --------------------------- pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # RMSE # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) # #create figure axes # fig, ax = plt.subplots(figsize = (10,10)) # #plot rms value # ax.bar(np.array([1,3,5])-0.3, df_sum_reg_stan.nerg_tot_rms.values, width=0.6, label='STAN') # ax.bar(np.array([1,3,5])+0.3, df_sum_reg_inla.nerg_tot_rms.values, width=0.6, label='INLA') # #figure properties # ax.set_ylim([0, 0.2]) # ax.set_xticks([1,3,5], df_sum_reg_stan.ds_name) # ax.set_xlabel('Dataset', fontsize=35) # ax.set_ylabel('RMSE', fontsize=35) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=32) # ax.tick_params(axis='y', labelsize=32) # ax.yaxis.set_major_formatter(FormatStrFormatter('%.2f')) # #legend # ax.legend(loc='upper left', fontsize=32) # #save figure # fig.tight_layout() # fig.savefig( fig_fname + '.png' ) # Run Time # --- --- --- --- --- #run time figure fig_fname = '%s/%s_run_time'%(dir_out, cmp_name) #create figure axes fig, ax = plt.subplots(figsize = (10,10)) #plot rms value ax.bar(np.array([1,3,5])-0.3, df_sum_reg_stan.run_time.values, width=0.6, label='STAN') ax.bar(np.array([1,3,5])+0.3, df_sum_reg_inla.run_time.values, width=0.6, label='INLA') #figure properties # ax.set_ylim([0, 0.2]) ax.set_xticks([1,3,5], df_sum_reg_stan.ds_name) ax.set_xlabel('Dataset', fontsize=35) ax.set_ylabel('Run Time (min)', fontsize=35) ax.grid(which='both') ax.set_yscale('log') ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) # ax.yaxis.set_major_formatter(FormatStrFormatter('%.2f')) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' )
5,403
31.95122
158
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/comparison_model1_misfit.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 15 14:50:27 2022 @author: glavrent """ # Working directory and Packages # --------------------------- #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt #user functions def PlotRSMCmp(df_rms_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for k in df_rms_all: df_rms = df_rms_all[k] ds_id = np.array(range(len(df_rms))) ax.plot(ds_id, df_rms.loc[:,c_name+'_rms'], linestyle='-', marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('RMSE', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_rms.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax def PlotKLCmp(df_KL_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for k in df_KL_all: df_KL = df_KL_all[k] ds_id = np.array(range(len(df_KL))) ax.plot(ds_id, df_KL.loc[:,c_name+'_KL'], linestyle='-', marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('KL divergence', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_KL.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax # Define variables # --------------------------- # COMPARISONS # Different Packages # --- --- --- --- --- cmp_name = 'STAN_pckg_cmp_NGAWest2CANorth' reg_title = ['PYSTAN2', 'PYSTAN3', 'CMDSTANPY'] reg_fname = ['PYSTAN_NGAWest2CANorth_chol_eff_small_corr_len','PYSTAN3_NGAWest2CANorth_chol_eff_small_corr_len','CMDSTAN_NGAWest2CANorth_chol_eff_small_corr_len'] ylim_time = [0, 700] # # Different Implementations # # --- --- --- --- --- # cmp_name = 'STAN_impl_cmp_NGAWest2CANorth' # reg_title = ['CMDSTANPY Chol.', 'CMDSTANPY Chol. Eff.'] # reg_fname = ['CMDSTAN_NGAWest2CANorth_chol_small_corr_len','CMDSTAN_NGAWest2CANorth_chol_eff_small_corr_len'] # ylim_time = [0, 700] # # Different Software # # --- --- --- --- --- # cmp_name = 'STAN_vs_INLA_cmp_NGAWest2CANorth' # reg_title = ['STAN','INLA'] # reg_fname = ['CMDSTAN_NGAWest2CANorth_chol_eff_small_corr_len','INLA_NGAWest2CANorth_coarse_small_corr_len'] # ylim_time = [0, 700] # Different # --- --- --- --- --- # # NGAWest2CANorth # cmp_name = 'INLA_mesh_cmp_NGAWest2CANorth' # reg_title = ['INLA coarse mesh', 'INLA medium mesh', 'INLA fine mesh'] # reg_fname = ['INLA_NGAWest2CANorth_coarse_small_corr_len','INLA_NGAWest2CANorth_medium_small_corr_len','INLA_NGAWest2CANorth_fine_small_corr_len'] # ylim_time = [0, 20] # # NGAWest2CANorth # cmp_name = 'INLA_mesh_cmp_NGAWest3CA' # reg_title = ['INLA coarse mesh', 'INLA medium mesh', 'INLA fine mesh'] # reg_fname = ['INLA_NGAWest3CA_coarse_small_corr_len','INLA_NGAWest3CA_medium_small_corr_len','INLA_NGAWest3CA_fine_small_corr_len'] # ylim_time = [0, 100] #directories regressions reg_dir = [f'../../../../Data/Verification/regression/ds1/%s/'%r_f for r_f in reg_fname] #directory output dir_out = '../../../../Data/Verification/regression/ds1/comparisons/' # Load Data # --------------------------- #initialize misfit dataframe df_sum_misfit_all = {}; #read misfit info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename misfit info fname_sum = r_d + 'summary/misfit_summary.csv' #read KL score for coefficients df_sum_misfit_all[r_t] = pd.read_csv(fname_sum, index_col=0) #initialize run time dataframe df_runinfo_all = {}; #read run time info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename run time fname_runinfo = r_d + '/run_info.csv' #store calc time df_runinfo_all[r_t] = pd.read_csv(fname_runinfo) # Comparison Figures # --------------------------- pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # RMSE divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); # KL divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); # Run Time # --- --- --- --- --- #run time figure fig_fname = '%s/%s_run_time'%(dir_out, cmp_name) #create figure axes fig, ax = plt.subplots(figsize = (10,10)) #iterate over different analyses for j, k in enumerate(df_runinfo_all): ds_id = df_runinfo_all[k].ds_id ds_name = ['Y%i'%d_i for d_i in ds_id] run_time = df_runinfo_all[k].run_time ax.plot(ds_id, run_time, marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim(ylim_time) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('Run Time (min)', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=ds_name) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='lower left', fontsize=32) # ax.legend(loc='upper left', fontsize=32) # ax.legend(loc='center left', bbox_to_anchor=(1, 0.5), fontsize=25) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' )
7,051
29.79476
162
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/comparison_inla_model1.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Aug 12 20:52:09 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load variables import os import sys import pathlib import glob import re #regular expression package import pickle #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick #user functions sys.path.insert(0,'../../../Python_lib/regression/') from pylib_stats import CalcRMS from pylib_stats import CalcLKDivergece # Define variables # --------------------------- # USER SETS DIRECTORIES AND FILE INFO OF SYNTHETIC DS AND REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #processed dataset # name_dataset = 'NGAWest2CANorth' # name_dataset = 'NGAWest2CA' # name_dataset = 'NGAWest3CA' #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts # corr_id = 1 #kernel function # 1: Mattern kernel (alpha=2) # 2: Negative Exp (alpha=3/2) # ker_id = 1 #mesh type # 1: Fine Mesh # 2: Medium Mesh # 3: Coarse Mesh # mesh_id = 3 #directories (synthetic dataset) if corr_id == 1: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds1_small_corr_len' elif corr_id == 2: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds1_large_corr_len' #directories (regression results) if mesh_id == 1: dir_results = f'../../../../Data/Verification/regression/ds1/INLA_%s_fine'%name_dataset elif mesh_id == 2: dir_results = f'../../../../Data/Verification/regression/ds1/INLA_%s_medium'%name_dataset elif mesh_id == 3: dir_results = f'../../../../Data/Verification/regression/ds1/INLA_%s_coarse'%name_dataset #prefix for synthetic data and results prfx_syndata = 'CatalogNGAWest3CALite_synthetic' #regression results filename prefix prfx_results = f'%s_syndata'%name_dataset # dataset info ds_id = np.arange(1,6) # ++++++++++++++++++++++++++++++++++++++++ # USER NEEDS TO SPECIFY HYPERPARAMETERS OF SYNTHETIC DATASET # ++++++++++++++++++++++++++++++++++++++++ # hyper-parameters if corr_id == 1: # small correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, 'ell_1e':60, 'ell_1as':30, 'phi_0':0.4, 'tau_0':0.3 } elif corr_id == 2: #large correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, 'ell_1e':100, 'ell_1as':70, 'phi_0':0.4, 'tau_0':0.3 } # ++++++++++++++++++++++++++++++++++++++++ # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #output filename sufix if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #kenel info if ker_id == 1: ker_suffix = '' elif ker_id == 2: ker_suffix = '_nexp' # ++++++++++++++++++++++++++++++++++++++++ #ploting options flag_report = True # Compare coefficients # --------------------------- #initialize misfit metrics dataframe df_misfit = pd.DataFrame(index=['Y%i'%d_id for d_id in ds_id]) #iterate over different datasets for d_id in ds_id: # Load Data #--- --- --- --- --- #file names #synthetic data fname_sdata_gmotion = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'data', synds_suffix, d_id) + '.csv' #regression results fname_reg_gmotion = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id, prfx_results, synds_suffix, d_id, 'residuals') + '.csv' fname_reg_coeff = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id, prfx_results, synds_suffix, d_id, 'coefficients') + '.csv' #load synthetic results df_sdata_gmotion = pd.read_csv(fname_sdata_gmotion).set_index('rsn') #load regression results df_reg_gmotion = pd.read_csv(fname_reg_gmotion, index_col=0) df_reg_coeff = pd.read_csv(fname_reg_coeff, index_col=0) # Processing #--- --- --- --- --- #keep only common records from synthetic dataset df_sdata_gmotion = df_sdata_gmotion.reindex(df_reg_gmotion.index) #find unique earthqakes and stations eq_id, eq_idx, eq_nrec = np.unique(df_sdata_gmotion.eqid, return_index=True, return_counts=True) sta_id, sta_idx, sta_nrec = np.unique(df_sdata_gmotion.ssn, return_index=True, return_counts=True) # Compute Root Mean Square Error #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_rms'] = CalcRMS(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_rms'] = CalcRMS(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_rms'] = CalcRMS(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_rms'] = CalcRMS(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) # Compute Divergence #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_KL'] = CalcLKDivergece(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) # Output #--- --- --- --- --- #figure directory dir_fig = '%s%s/Y%i/figures_cmp/'%(dir_results,ker_suffix+synds_suffix,d_id) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #compare ground motion predictions #... ... ... ... ... ... #figure title fname_fig = 'Y%i_tot_res_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #median ax.scatter(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title('Comparison total residuals, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2,2]) ax.set_ylim([-2,2]) fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1e #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1e_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,e}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-.4,.4]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['dc_1e_sig'].values[eq_idx], df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,e}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,e}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1as #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1as_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1as_sig'].values[sta_idx], df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1bs #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1bs_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1bs_sig'].values[sta_idx], df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,s}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare Misfit Metrics # --------------------------- #summary directory dir_sum = '%s%s/summary/'%(dir_results,ker_suffix+synds_suffix) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #figure directory dir_fig = '%s/figures/'%(dir_sum) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save df_misfit.to_csv(dir_sum + 'misfit_summary.csv') #RMS misfit fname_fig = 'misfit_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,e}$') ax.plot(ds_id, df_misfit.dc_1as_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,s}$') ax.plot(ds_id, df_misfit.dc_1bs_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,s}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('RSME', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #KL divergence fname_fig = 'KLdiv_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,e}$') ax.plot(ds_id, df_misfit.dc_1as_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,s}$') ax.plot(ds_id, df_misfit.dc_1bs_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,s}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('KL divergence', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare hyper-paramters # --------------------------- #iterate over different datasets df_reg_hyp = list() df_reg_hyp_post = list() for d_id in ds_id: # Load Data #--- --- --- --- --- #regression hyperparamters results fname_reg_hyp = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results,synds_suffix, d_id,prfx_results, synds_suffix, d_id, 'hyperparameters') + '.csv' fname_reg_hyp_post = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results,synds_suffix, d_id,prfx_results, synds_suffix, d_id, 'hyperposterior') + '.csv' #load regression results df_reg_hyp.append( pd.read_csv(fname_reg_hyp, index_col=0) ) df_reg_hyp_post.append( pd.read_csv(fname_reg_hyp_post, index_col=0) ) # Omega_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mode = 40 ymax_mean = 40 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1,e}$', fontsize=30) ax.set_xlabel('$\omega_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.25]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mode = 30 ymax_mean = 30 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1a,s}$', fontsize=30) ax.set_xlabel('$\omega_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1bs #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1bs' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 60 # ymax_mean = 60 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1b,s}$', fontsize=30) ax.set_xlabel('$\omega_{1b,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mode = 0.02 ymax_mean = 0.02 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'0.5quant'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1,e}$', fontsize=30) ax.set_xlabel('$\ell_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1a,s}$', fontsize=30) ax.set_xlabel('$\ell_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Tau_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'tau_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\tau_{0}$', fontsize=30) ax.set_xlabel(r'$\tau_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Phi_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'phi_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mode = 100 ymax_mean = 100 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\phi_{0}$', fontsize=30) ax.set_xlabel('$\phi_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.6]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # # Delta c_0 # #--- --- --- --- --- # #hyper-paramter name # name_hyp = 'dc_0' # #figure title # fname_fig = 'post_dist_' + name_hyp # #create figure # fig, ax = plt.subplots(figsize = (10,10)) # for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): # #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mean = 15 # #plot posterior dist # pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') # #plot true value # ymax_hyp = ymax_mean # # ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') # #edit figure # ax.set_title(r'Comparison $\delta c_{0}$', fontsize=30) # ax.set_xlabel('$\delta c_{0}$', fontsize=25) # ax.set_ylabel('probability density function ', fontsize=25) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=22) # ax.tick_params(axis='y', labelsize=22) # #plot limits # ax.set_xlim([-1,1]) # ax.set_ylim([0,ymax_hyp]) # #save figure # fig.tight_layout() # # fig.savefig( dir_fig + fname_fig + '.png' )
30,341
38.71466
155
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/main_cmdstan_model1_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') from regression_cmdstan_model1_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds1' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info # ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #stan model # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds1/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest3CA' # out_dir_sub = 'CMDSTAN_NGAWest3CA_chol' # out_dir_sub = 'CMDSTAN_NGAWest3CA_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest3CA_chol_eff2' #stan parameters res_name='tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #parallel options stan_parallel=False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, sm_fname, out_fname, out_dir, res_name, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=stan_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,227
29.742857
92
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/comparison_inla_model1_misfit_mesh.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 15 14:50:27 2022 @author: glavrent """ # Working directory and Packages # --------------------------- #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt #user functions def PlotRSMCmp(df_rms_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) ltype_array = ['-','--',':'] for j, k in enumerate(df_rms_all): df_rms = df_rms_all[k] ds_id = np.array(range(len(df_rms))) #plot info lcol = mpl.cm.get_cmap('tab10')( np.floor_divide(j,3) ) ltype = ltype_array[ np.mod(j,3) ] ax.plot(ds_id, df_rms.loc[:,c_name+'_rms'], marker='o', linewidth=2, markersize=10, label=k, linestyle=ltype, color=lcol) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('RMSE', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_rms.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend # ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax def PlotKLCmp(df_KL_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for k in df_KL_all: df_KL = df_KL_all[k] ds_id = np.array(range(len(df_KL))) ax.plot(ds_id, df_KL.loc[:,c_name+'_KL'], linestyle='-', marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('KL divergence', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_KL.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend # ax.legend(loc='upper left', fontsize=32) # ax.legend(loc='center left', bbox_to_anchor=(1, 0.5), fontsize=25) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax # Define variables # --------------------------- # COMPARISONS # Different Mesh sizes # --- --- --- --- --- cmp_name = 'INLA_mesh' reg_title = [f'NGAW3* CA\nfine', f'NGAW3* CA \nmedium', f'NGAW3 CA \ncoarse', f'NGAW2 CA \nfine', f'NGAW2 CA \nmedium', f'NGAW2 CA \ncoarse', f'NGAW2 CA, North\nfine', f'NGAW2 CA, North\nmedium', f'NGAW2 CA, North\ncoarse'] reg_fname = ['INLA_NGAWest3CA_fine_small_corr_len', 'INLA_NGAWest3CA_medium_small_corr_len', 'INLA_NGAWest3CA_coarse_small_corr_len', 'INLA_NGAWest2CA_fine_small_corr_len', 'INLA_NGAWest2CA_medium_small_corr_len', 'INLA_NGAWest2CA_coarse_small_corr_len', 'INLA_NGAWest2CANorth_fine_small_corr_len','INLA_NGAWest2CANorth_medium_small_corr_len','INLA_NGAWest2CANorth_coarse_small_corr_len'] ylim_time = [0, 50] # # Different Implementations # # --- --- --- --- --- # cmp_name = 'STAN_impl_cmp_NGAWest2CANorth' # reg_title = ['CMDSTANPY Chol.', 'CMDSTANPY Chol. Eff.'] # reg_fname = ['CMDSTAN_NGAWest2CANorth_chol_small_corr_len','CMDSTAN_NGAWest2CANorth_chol_eff_small_corr_len'] # ylim_time = [0, 700] # # Different Software # # --- --- --- --- --- # cmp_name = 'STAN_vs_INLA_cmp_NGAWest2CANorth' # reg_title = ['STAN','INLA'] # reg_fname = ['CMDSTAN_NGAWest2CANorth_chol_eff_small_corr_len','INLA_NGAWest2CANorth_coarse_small_corr_len'] # ylim_time = [0, 700] # Different # --- --- --- --- --- # # NGAWest2CANorth # cmp_name = 'INLA_mesh_cmp_NGAWest2CANorth' # reg_title = ['INLA coarse mesh', 'INLA medium mesh', 'INLA fine mesh'] # reg_fname = ['INLA_NGAWest2CANorth_coarse_small_corr_len','INLA_NGAWest2CANorth_medium_small_corr_len','INLA_NGAWest2CANorth_fine_small_corr_len'] # ylim_time = [0, 20] # # NGAWest2CANorth # cmp_name = 'INLA_mesh_cmp_NGAWest3CA' # reg_title = ['INLA coarse mesh', 'INLA medium mesh', 'INLA fine mesh'] # reg_fname = ['INLA_NGAWest3CA_coarse_small_corr_len','INLA_NGAWest3CA_medium_small_corr_len','INLA_NGAWest3CA_fine_small_corr_len'] # ylim_time = [0, 100] #directories regressions reg_dir = [f'../../../../Data/Verification/regression/ds1/%s/'%r_f for r_f in reg_fname] #directory output dir_out = '../../../../Data/Verification/regression/ds1/comparisons/' # Load Data # --------------------------- #initialize misfit dataframe df_sum_misfit_all = {}; #read misfit info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename misfit info fname_sum = r_d + 'summary/misfit_summary.csv' #read KL score for coefficients df_sum_misfit_all[r_t] = pd.read_csv(fname_sum, index_col=0) #initialize run time dataframe df_runinfo_all = {}; #read run time info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename run time fname_runinfo = r_d + '/run_info.csv' #store calc time df_runinfo_all[r_t] = pd.read_csv(fname_runinfo) # print(f'%s: %.1f min'%( r_t, df_runinfo_all[r_t].run_time.mean() )) # Comparison Figures # --------------------------- pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # RMSE divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); # KL divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); # Run Time # --- --- --- --- --- #run time figure fig_fname = '%s/%s_run_time'%(dir_out, cmp_name) #create figure axes # fig, ax = plt.subplots(figsize = (10,10)) fig, ax = plt.subplots(figsize = (14,10)) ltype_array = ['-','--',':'] #iterate over different analyses for j, k in enumerate(df_runinfo_all): ds_id = df_runinfo_all[k].ds_id ds_name = ['Y%i'%d_i for d_i in ds_id] run_time = df_runinfo_all[k].run_time lcol = mpl.cm.get_cmap('tab10')( np.floor_divide(j,3) ) ltype = ltype_array[ np.mod(j,3) ] ax.plot(ds_id, run_time, marker='o', linewidth=2, markersize=10, label=k, linestyle=ltype, color=lcol) #figure properties ax.set_ylim(ylim_time) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('Run Time (min)', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=ds_name) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend # ax.legend(loc='lower left', fontsize=32) # ax.legend(loc='upper left', fontsize=32) ax.legend(loc='center left', bbox_to_anchor=(1, 0.5), fontsize=25) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' )
8,074
32.094262
148
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/comparison_stan_model1_misfit.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 15 14:50:27 2022 @author: glavrent """ # Working directory and Packages # --------------------------- #change working directory import os os.chdir('/mnt/halcloud_nfs/glavrent/Research/Nonerg_GMM_methodology/Analyses/Code_Verification/regressions/ds1') #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt #user functions def PlotRSMCmp(df_rms_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for k in df_rms_all: df_rms = df_rms_all[k] ds_id = np.array(range(len(df_rms))) ax.plot(ds_id, df_rms.loc[:,c_name+'_rms'], linestyle='-', marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('RMSE', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_rms.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax def PlotKLCmp(df_KL_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for k in df_KL_all: df_KL = df_KL_all[k] ds_id = np.array(range(len(df_KL))) ax.plot(ds_id, df_KL.loc[:,c_name+'_KL'], linestyle='-', marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('KL divergence', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_KL.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax # Define variables # --------------------------- # COMPARISONS # # Different Packages # # --- --- --- --- --- # cmp_name = 'STAN_pckg_cmp_NGAWest2CANorth' # reg_title = ['PYSTAN2', 'PYSTAN3', 'CMDSTANPY'] # reg_fname = ['PYSTAN_NGAWest2CANorth_chol_eff_small_corr_len','PYSTAN3_NGAWest2CANorth_chol_eff_small_corr_len','CMDSTAN_NGAWest2CANorth_chol_eff_small_corr_len'] # ylim_time = [0, 700] # Different Implementations # --- --- --- --- --- cmp_name = 'STAN_impl_cmp_NGAWest2CANorth' reg_title = ['CMDSTANPY Chol.', 'CMDSTANPY Chol. Eff.'] reg_fname = ['CMDSTAN_NGAWest2CANorth_chol_small_corr_len','CMDSTAN_NGAWest2CANorth_chol_eff_small_corr_len'] # reg_fname = ['PYSTAN_NGAWest2CANorth_chol_small_corr_len','PYSTAN_NGAWest2CANorth_chol_eff_small_corr_len'] ylim_time = [0, 700] #directories regressions reg_dir = [f'../../../../Data/Verification/regression/ds1/%s/'%r_f for r_f in reg_fname] #directory output dir_out = '../../../../Data/Verification/regression/ds1/comparisons/' # Load Data # --------------------------- #initialize misfit dataframe df_sum_misfit_all = {}; #read misfit info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename misfit info fname_sum = r_d + 'summary/misfit_summary.csv' #read KL score for coefficients df_sum_misfit_all[r_t] = pd.read_csv(fname_sum, index_col=0) #initialize run time dataframe df_runinfo_all = {}; #read run time info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename run time fname_runinfo = r_d + '/run_info.csv' #store calc time df_runinfo_all[r_t] = pd.read_csv(fname_runinfo) # Comparison Figures # --------------------------- pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # RMSE divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); # KL divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); # RMSE divergence # --- --- --- --- --- #run time figure fig_fname = '%s/%s_run_time'%(dir_out, cmp_name) #create figure axes fig, ax = plt.subplots(figsize = (10,10)) #iterate over different analyses for j, k in enumerate(df_runinfo_all): ds_id = df_runinfo_all[k].ds_id ds_name = ['Y%i'%d_i for d_i in ds_id] run_time = df_runinfo_all[k].run_time ax.plot(ds_id, run_time, marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim(ylim_time) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('Run Time (min)', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=ds_name) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='lower left', fontsize=32) # ax.legend(loc='center left', bbox_to_anchor=(1, 0.5), fontsize=25) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' )
6,363
28.738318
164
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/main_pystan_model1_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model1_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds1' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info # ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #stan model # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds1/' #output sub-directory #python 2 # out_dir_sub = 'PYSTAN_NGAWest2CA' # out_dir_sub = 'PYSTAN_NGAWest2CA_chol' # out_dir_sub = 'PYSTAN_NGAWest2CA_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest2CA_chol_eff2' #python 3 # out_dir_sub = 'PYSTAN3_NGAWest2CA' # out_dir_sub = 'PYSTAN3_NGAWest2CA_chol' # out_dir_sub = 'PYSTAN3_NGAWest2CA_chol_eff' # out_dir_sub = 'PYSTAN3_NGAWest2CA_chol_eff2' #stan parameters runstan_flag = True #pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, sm_fname, out_fname, out_dir, res_name, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,544
29.826087
90
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/main_cmdstan_model1_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') from regression_cmdstan_model1_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds1' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info # ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #stan model # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds1/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CANorth' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_chol' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_chol_eff2' #stan parameters res_name='tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #parallel options stan_parallel=False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, sm_fname, out_fname, out_dir, res_name, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=stan_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,436
30.824074
92
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/main_cmdstan_model1_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') from regression_cmdstan_model1_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds1' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info # ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #stan model # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds1/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CA' # out_dir_sub = 'CMDSTAN_NGAWest2CA_chol' # out_dir_sub = 'CMDSTAN_NGAWest2CA_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CA_chol_eff2' #stan parameters res_name='tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #parallel options stan_parallel=False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, sm_fname, out_fname, out_dir, res_name, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=stan_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,316
30
92
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/main_pystan_model1_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model1_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds1' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #stan model # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model1_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds1/' #output sub-directory # out_dir_sub = 'PYSTAN_NGAWest3CA' # out_dir_sub = 'PYSTAN_NGAWest3CA_chol' # out_dir_sub = 'PYSTAN_NGAWest3CA_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest3CA_chol_eff2' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, sm_fname, out_fname, out_dir, res_name, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,244
29.327103
90
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds1/comparison_inla_model1_misfit.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 15 14:50:27 2022 @author: glavrent """ # Working directory and Packages # --------------------------- #change working directory import os os.chdir('/mnt/halcloud_nfs/glavrent/Research/Nonerg_GMM_methodology/Analyses/Code_Verification/regressions/ds1') #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt #user functions def PlotRSMCmp(df_KL_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for m_i in df_KL_all: df_KL = df_KL_all[m_i] ds_id = np.array(range(len(df_KL))) ax.plot(ds_id, df_KL.loc[:,c_name], linestyle='-', marker='o', linewidth=2, markersize=10, label=m_i) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=30) ax.set_ylabel('RMSE divergence', fontsize=30) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_KL.index) ax.tick_params(axis='x', labelsize=30) ax.tick_params(axis='y', labelsize=30) #legend ax.legend(loc='upper left', fontsize=30) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax def PlotKLCmp(df_KL_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for m_i in df_KL_all: df_KL = df_KL_all[m_i] ds_id = np.array(range(len(df_KL))) ax.plot(ds_id, df_KL.loc[:,c_name], linestyle='-', marker='o', linewidth=2, markersize=10, label=m_i) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=30) ax.set_ylabel('KL divergence', fontsize=30) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_KL.index) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax # Define variables # --------------------------- #comparisons name_reg = ['PYSTAN_NGAWest2CANorth_chol_eff_small_corr_len','PYSTAN3_NGAWest2CANorth_chol_eff_small_corr_len',] #dataset info ds_id = 1 #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #packages comparison packg_info = ['PYSTAN', 'PYSTAN3', 'fine'] #correlation name if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #dataset name if ds_id == 1: name_dataset = 'NGAWest2CANorth' elif ds_id == 2: name_dataset = 'NGAWest2CA' elif ds_id == 3: name_dataset = 'NGAWest3CA' #directories regressions dir_reg = [f'../../../../Data/Verification/regression/ds1/%s_%s_%s%s/'%(name_dataset, m_info, synds_suffix) for m_info in mesh_info] #directory output dir_out = '../../../../Data/Verification/regression/ds1/comparisons/' # Load Data # --------------------------- #initialize dataframe df_KL_all = {}; #read KL scores for k, (d_r, m_i) in enumerate(zip(dir_reg, mesh_info)): #filename KL score fname_KL = d_r + 'summary/coeffs_KL_divergence.csv' #read KL score for coefficients df_KL_all[m_i] = pd.read_csv(fname_KL, index_col=0) # Comparison Figures # --------------------------- pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s%s_KLdiv_%s'%(dir_out, name_dataset, synds_suffix, c_name) #plotting PlotKLCmp(df_KL_all, c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s%s_KLdiv_%s'%(dir_out, name_dataset, synds_suffix, c_name) #plotting PlotKLCmp(df_KL_all, c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s%s_KLdiv_%s'%(dir_out, name_dataset, synds_suffix, c_name) #plotting PlotKLCmp(df_KL_all, c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s%s_KLdiv_%s'%(dir_out, name_dataset, synds_suffix, c_name) #plotting PlotKLCmp(df_KL_all, c_name, fig_fname);
4,303
26.767742
132
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_corr_cells_NGAWest2CANorth_sparse.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory out_dir_sub = 'PYSTAN_NGAWest2CANorth_corr_cells_chol_eff_sp' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,573
29.547009
105
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_corr_cells_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_corr_cells_unbounded_hyp import RunStan # from regression_pystan_model2_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient2.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory #pystan 2 # out_dir_sub = 'PYSTAN_NGAWest2CANorth_corr_cells' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_corr_cells_chol' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_corr_cells_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_corr_cells_chol_eff2' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_corr_cells_chol_eff_sp' #pystan 3 # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_corr_cells' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_corr_cells_chol' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_corr_cells_chol_eff' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_corr_cells_chol_eff2' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_corr_cells_chol_eff_sp' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,608
33.916667
107
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/comparison_stan_model2_corr_cells.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Aug 12 10:26:06 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load packages import os import sys import pathlib import glob import re #regular expression package import pickle #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick #user functions sys.path.insert(0,'../../../Python_lib/regression/') from pylib_stats import CalcRMS from pylib_stats import CalcLKDivergece # Define variables # --------------------------- # USER SETS DIRECTORIES AND FILE INFO OF SYNTHETIC DS AND REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #processed dataset # name_dataset = 'NGAWest2CANorth' name_dataset = 'NGAWest2CA' # name_dataset = 'NGAWest3CA' #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #package # 1: Pystan v2 # 2: Pystan v3 # 3: stancmd pkg_id = 3 #approximation type # 1: multivariate normal # 2: cholesky # 3: cholesky efficient # 4: cholesky efficient v2 # 5: cholesky efficient, sparse cells aprox_id = 5 #directories (synthetic dataset) if corr_id == 1: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds2_small_corr_len' elif corr_id == 2: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds2_large_corr_len' #cell info fname_cellinfo = dir_syndata + '/' + 'CatalogNGAWest3CALite_cellinfo.csv' fname_distmat = dir_syndata + '/' + 'CatalogNGAWest3CALite_distancematrix.csv' #directories (regression results) if pkg_id == 1: dir_results = f'../../../../Data/Verification/regression/ds2/PYSTAN_%s'%name_dataset elif pkg_id == 2: dir_results = f'../../../../Data/Verification/regression/ds2/PYSTAN3_%s'%name_dataset elif pkg_id == 3: dir_results = f'../../../../Data/Verification/regression/ds2/CMDSTAN_%s'%name_dataset #directories (regression results) if pkg_id == 1: dir_results = f'../../../../Data/Verification/regression_old/ds2/PYSTAN_%s'%name_dataset elif pkg_id == 2: dir_results = f'../../../../Data/Verification/regression_old/ds2/PYSTAN3_%s'%name_dataset elif pkg_id == 3: dir_results = f'../../../../Data/Verification/regression_old/ds2/CMDSTAN_%s'%name_dataset #prefix for synthetic data and results prfx_syndata = 'CatalogNGAWest3CALite_synthetic' #regression results filename prefix prfx_results = f'%s_syndata'%name_dataset # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #output filename sufix (synthetic dataset) if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #output filename sufix (regression results) if aprox_id == 1: synds_suffix_stan = '_corr_cells' + synds_suffix elif aprox_id == 2: synds_suffix_stan = '_corr_cells' + '_chol' + synds_suffix elif aprox_id == 3: synds_suffix_stan = '_corr_cells' + '_chol_eff' + synds_suffix elif aprox_id == 4: synds_suffix_stan = '_corr_cells' + '_chol_eff2' + synds_suffix elif aprox_id == 5: synds_suffix_stan = '_corr_cells' + '_chol_eff_sp' + synds_suffix # dataset info ds_id = np.arange(1,6) # ++++++++++++++++++++++++++++++++++++++++ # USER NEEDS TO SPECIFY HYPERPARAMETERS OF SYNTHETIC DATASET # ++++++++++++++++++++++++++++++++++++++++ # hyper-parameters if corr_id == 1: # small correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, 'ell_1e':60, 'ell_1as':30, 'c_cap_erg': -0.011, 'omega_cap_mu': 0.005, 'omega_ca1p':0.004, 'omega_ca2p':0.002, 'ell_ca1p': 75, 'phi_0':0.4, 'tau_0':0.3 } elif corr_id == 2: #large correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, 'ell_1e':100, 'ell_1as':70, 'c_cap_erg': -0.02, 'omega_cap_mu': 0.008, 'omega_ca1p':0.005, 'omega_ca2p':0.003, 'ell_ca1p': 120, 'phi_0':0.4, 'tau_0':0.3} # ++++++++++++++++++++++++++++++++++++++++ #ploting options flag_report = True # Compare results # --------------------------- #load cell data df_cellinfo = pd.read_csv(fname_cellinfo).set_index('cellid') df_distmat = pd.read_csv(fname_distmat).set_index('rsn') #initialize misfit metrics dataframe df_misfit = pd.DataFrame(index=['Y%i'%d_id for d_id in ds_id]) #iterate over different datasets for d_id in ds_id: # Load Data #--- --- --- --- --- #file names #synthetic data fname_sdata_gmotion = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'data', synds_suffix, d_id) + '.csv' fname_sdata_atten = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'atten', synds_suffix, d_id) + '.csv' #regression results fname_reg_gmotion = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'residuals') + '.csv' fname_reg_coeff = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'coefficients') + '.csv' fname_reg_atten = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'catten') + '.csv' #load synthetic results df_sdata_gmotion = pd.read_csv(fname_sdata_gmotion).set_index('rsn') df_sdata_atten = pd.read_csv(fname_sdata_atten).set_index('cellid') #load regression results df_reg_gmotion = pd.read_csv(fname_reg_gmotion, index_col=0) df_reg_coeff = pd.read_csv(fname_reg_coeff, index_col=0) df_reg_atten = pd.read_csv(fname_reg_atten, index_col=0) # Processing #--- --- --- --- --- #keep only relevant columns from synthetic dataset df_sdata_gmotion = df_sdata_gmotion.reindex(df_reg_gmotion.index) df_sdata_atten = df_sdata_atten.reindex(df_reg_atten.index) #distance matrix for records of interest df_dmat = df_distmat.reindex(df_sdata_gmotion.index) #find unique earthqakes and stations eq_id, eq_idx, eq_nrec = np.unique(df_sdata_gmotion.eqid, return_index=True, return_counts=True) sta_id, sta_idx, sta_nrec = np.unique(df_sdata_gmotion.ssn, return_index=True, return_counts=True) #number of paths per cell cell_npath = np.sum(df_dmat.loc[:,df_reg_atten.cellname] > 0, axis=0) # Compute Root Mean Square Error #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_rms'] = CalcRMS(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_rms'] = CalcRMS(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_rms'] = CalcRMS(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_rms'] = CalcRMS(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_rms'] = CalcRMS(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Compute Divergence #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_KL'] = CalcLKDivergece(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_KL'] = CalcLKDivergece(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Output #--- --- --- --- --- #figure directory dir_fig = '%s%s/Y%i/figures_cmp/'%(dir_results,synds_suffix_stan,d_id) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #compare ground motion predictions #... ... ... ... ... ... #figure title fname_fig = 'Y%i_scatter_tot_res'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #median ax.scatter(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title('Comparison total residuals, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=35) ax.set_ylabel('Estimated', fontsize=35) ax.grid(which='both') ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-10,2]) ax.set_ylim([-10,2]) fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1e #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1e_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-.4,.4]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['dc_1e_sig'].values[eq_idx], df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1as #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1as_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1as_sig'].values[sta_idx], df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1bs #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1bs_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1bs_sig'].values[sta_idx], df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_cap #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_cap_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-0.05,0.02]) ax.set_ylim([-0.05,0.02]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_atten['c_cap_sig'], df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.00,0.03]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_npath'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(cell_npath, df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of paths', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,5e4]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare Misfit Metrics # --------------------------- #summary directory dir_sum = '%s%s/summary/'%(dir_results,synds_suffix_stan) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #figure directory dir_fig = '%s/figures/'%(dir_sum) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save df_misfit.to_csv(dir_sum + 'misfit_summary.csv') #RMS misfit fname_fig = 'misfit_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_cap_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('RSME', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #KL divergence fname_fig = 'KLdiv_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_cap_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('KL divergence', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare hyper-paramters # --------------------------- #iterate over different datasets df_reg_hyp = list() df_reg_hyp_post = list() for d_id in ds_id: # Load Data #--- --- --- --- --- #regression hyperparamters results fname_reg_hyp = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperparameters') + '.csv' fname_reg_hyp_post = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperposterior') + '.csv' #load regression results df_reg_hyp.append( pd.read_csv(fname_reg_hyp, index_col=0) ) df_reg_hyp_post.append( pd.read_csv(fname_reg_hyp_post, index_col=0) ) # Omega_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 40 ymax_mean = 40 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1,E}$', fontsize=30) ax.set_xlabel('$\omega_{1,E}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.25]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 30 ymax_mean = 30 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1a,S}$', fontsize=30) ax.set_xlabel('$\omega_{1a,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1bs #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1bs' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1b,S}$', fontsize=30) ax.set_xlabel('$\omega_{1b,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.02 ymax_mean = 0.02 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1,E}$', fontsize=30) ax.set_xlabel('$\ell_{1,E}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1a,S}$', fontsize=30) ax.set_xlabel('$\ell_{1a,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Tau_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'tau_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 150 ymax_mean = 150 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\tau_{0}$', fontsize=30) ax.set_xlabel(r'$\tau_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Phi_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'phi_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1000 ymax_mean = 1000 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\phi_{0}$', fontsize=30) ax.set_xlabel('$\phi_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.6]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_ca1 #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_ca1p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1500 ymax_mean = 1500 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp['omega_ca2p'], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{ca1,P}$', fontsize=30) ax.set_xlabel('$\omega_{ca1,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.05]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_ca2 #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_ca2p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1500 ymax_mean = 1500 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp['omega_ca2p'], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{ca2,P}$', fontsize=30) ax.set_xlabel('$\omega_{ca2,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.05]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_ca1p #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_ca1p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.02 ymax_mean = 0.02 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{ca1,P}$', fontsize=30) ax.set_xlabel('$\ell_{ca1,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # # Delta c_0 # #--- --- --- --- --- # #hyper-paramter name # name_hyp = 'dc_0' # #figure title # fname_fig = 'post_dist_' + name_hyp # #create figure # fig, ax = plt.subplots(figsize = (10,10)) # for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): # #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mean = 15 # #plot posterior dist # pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') # #plot true value # ymax_hyp = ymax_mean # # ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') # #edit figure # ax.set_title(r'Comparison $\delta c_{0}$', fontsize=30) # ax.set_xlabel('$\delta c_{0}$', fontsize=25) # ax.set_ylabel('probability density function ', fontsize=25) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=22) # ax.tick_params(axis='y', labelsize=22) # #plot limits # ax.set_xlim([-1,1]) # ax.set_ylim([0,ymax_hyp]) # #save figure # fig.tight_layout() # # fig.savefig( dir_fig + fname_fig + '.png' )
37,471
38.320042
153
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_corr_cells_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_corr_cells_unbounded_hyp import RunStan # from regression_pystan_model2_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient2.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_sparse_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory #python 2 # out_dir_sub = 'PYSTAN_NGAWest2CA_corr_cells' # out_dir_sub = 'PYSTAN_NGAWest2CA_corr_cells_chol' # out_dir_sub = 'PYSTAN_NGAWest2CA_corr_cells_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest2CA_corr_cells_chol_eff2' #python 3 # out_dir_sub = 'PYSTAN3_NGAWest2CA_corr_cells' # out_dir_sub = 'PYSTAN3_NGAWest2CA_corr_cells_chol' # out_dir_sub = 'PYSTAN3_NGAWest2CA_corr_cells_chol_eff' # out_dir_sub = 'PYSTAN3_NGAWest2CA_corr_cells_chol_eff2' # out_dir_sub = 'PYSTAN3_NGAWest2CA_corr_cells_chol_eff_sp' #stan parameters runstan_flag = True pystan_ver = 2 # pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 #0.9 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,412
32.687023
108
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/comparison_model2_misfit_stan_sparse.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 15 14:50:27 2022 @author: glavrent """ # Working directory and Packages # --------------------------- #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt #user functions def PlotRSMCmp(df_rms_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for j, k in enumerate(df_rms_all): df_rms = df_rms_all[k] ds_id = np.array(range(len(df_rms))) lcol = mpl.cm.get_cmap('tab10')(0) if j in [0,2] else mpl.cm.get_cmap('tab10')(1) ltype = '-' if j in [0,1] else '--' ax.plot(ds_id, df_rms.loc[:,c_name+'_rms'], marker='o', linewidth=2, markersize=10, label=k, linestyle=ltype, color=lcol) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('RMSE', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_rms.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax def PlotKLCmp(df_KL_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for k in df_KL_all: df_KL = df_KL_all[k] ds_id = np.array(range(len(df_KL))) ax.plot(ds_id, df_KL.loc[:,c_name+'_KL'], linestyle='-', marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('KL divergence', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_KL.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax # Define variables # --------------------------- # # Sparse Distance Matrix # # --- --- --- --- --- # # NGAWest 2 CA North # cmp_name = 'STAN_sparse_cmp_NGAWest2CA' # reg_title = ['STAN','STAN w/ sp dist matrix'] # reg_fname = ['CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len','CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_sp_small_corr_len'] # ylim_time = [0, 800] # NGAWest 2 CA cmp_name = 'STAN_sparse_cmp_NGAWest2CA' reg_title = ['STAN','STAN w/ sp dist matrix'] reg_fname = ['CMDSTAN_NGAWest2CA_corr_cells_chol_eff_small_corr_len','CMDSTAN_NGAWest2CA_corr_cells_chol_eff_sp_small_corr_len'] ylim_time = [0, 7000] # NGAWest 2 CA & NGAWest 2 CA North cmp_name = 'STAN_sparse_cmp_NGAWest2CA_' reg_title = ['STAN - NGAW2 CA','STAN - NGAW2 CA North', 'STAN - NGAW2 CA\nw/ sp dist matrix',f'STAN NGAW2 CA North\nw/ sp dist matrix, '] reg_fname = ['CMDSTAN_NGAWest2CA_corr_cells_chol_eff_small_corr_len', 'CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len', 'CMDSTAN_NGAWest2CA_corr_cells_chol_eff_sp_small_corr_len', 'CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_sp_small_corr_len'] ylim_time = [0, 7000] # # Different Software # # --- --- --- --- --- # cmp_name = 'STAN_vs_INLA_cmp_NGAWest2CANorth' # reg_title = ['STAN corr. cells','STAN uncorr. cells','INLA uncorr. cells'] # reg_fname = ['CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len','CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len', # 'INLA_NGAWest2CANorth_uncorr_cells_coarse_small_corr_len'] # reg_fname = ['PYSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len','PYSTAN_NGAWest2CANorth_uncorr_cells_chol_eff_small_corr_len', # 'INLA_NGAWest2CANorth_uncorr_cells_coarse_small_corr_len'] # ylim_time = [0, 800] #directories regressions reg_dir = [f'../../../../Data/Verification/regression/ds2/%s/'%r_f for r_f in reg_fname] #directory output dir_out = '../../../../Data/Verification/regression/ds2/comparisons/' # Load Data # --------------------------- #initialize misfit dataframe df_sum_misfit_all = {}; #read misfit info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename misfit info fname_sum = r_d + 'summary/misfit_summary.csv' #read KL score for coefficients df_sum_misfit_all[r_t] = pd.read_csv(fname_sum, index_col=0) #initialize run time dataframe df_runinfo_all = {}; #read run time info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename run time fname_runinfo = r_d + '/run_info.csv' #store calc time df_runinfo_all[r_t] = pd.read_csv(fname_runinfo) # Comparison Figures # --------------------------- pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # RMSE divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); # KL divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); # Run Time # --- --- --- --- --- #run time figure fig_fname = '%s/%s_run_time'%(dir_out, cmp_name) #create figure axes fig, ax = plt.subplots(figsize = (10,10)) #iterate over different analyses for j, k in enumerate(df_runinfo_all): ds_id = df_runinfo_all[k].ds_id ds_name = ['Y%i'%d_i for d_i in ds_id] run_time = df_runinfo_all[k].run_time # lcol = mpl.cm.get_cmap('tab10')(0) if j in [0,2] else mpl.cm.get_cmap('tab10')(1) ltype = '-' if j in [0,1] else '--' ax.plot(ds_id, run_time, marker='o', linewidth=2, markersize=10, label=k, linestyle=ltype, color=lcol) #figure properties ax.set_ylim(ylim_time) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('Run Time (min)', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=ds_name) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend # ax.legend(loc='lower left', fontsize=32) # ax.legend(loc='upper left', fontsize=32) # ax.legend(loc='center left', bbox_to_anchor=(1, 0.5), fontsize=25) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' )
7,569
31.48927
140
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_uncorr_cells_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_uncorr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory #pystan2 # out_dir_sub = 'PYSTAN_NGAWest2CANorth_uncorr_cells' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_uncorr_cells_chol' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_uncorr_cells_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest2CANorth_uncorr_cells_chol_eff2' #pystan3 # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_uncorr_cells' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_uncorr_cells_chol' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_uncorr_cells_chol_eff' # out_dir_sub = 'PYSTAN3_NGAWest2CANorth_uncorr_cells_chol_eff2' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,311
32.169231
103
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_corr_cells_NGAWest2CA_sparse.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory # out_dir_sub = 'PYSTAN_NGAWest2CA_corr_cells_chol_eff_sp' out_dir_sub = 'PYSTAN3_NGAWest2CA_corr_cells_chol_eff_sp' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 #0.9 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,545
29.307692
105
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/comparison_inla_model2_uncorr_cells.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Aug 12 10:26:06 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load packages import os import sys import pathlib import glob import re #regular expression package import pickle #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick #user functions sys.path.insert(0,'../../../Python_lib/regression/') from pylib_stats import CalcRMS from pylib_stats import CalcLKDivergece # Define variables # --------------------------- # USER SETS DIRECTORIES AND FILE INFO OF SYNTHETIC DS AND REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #processed dataset # name_dataset = 'NGAWest2CANorth' # name_dataset = 'NGAWest2CA' # name_dataset = 'NGAWest3CA' #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #kernel function # 1: Mattern kernel (alpha=2) # 2: Negative Exp (alpha=3/2) ker_id = 1 #mesh type # 1: Fine Mesh # 2: Medium Mesh # 3: Coarse Mesh mesh_id = 1 #directories (synthetic dataset) if corr_id == 1: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds2_small_corr_len' elif corr_id == 2: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds2_large_corr_len' #directories (regression results) if mesh_id == 1: dir_results = f'../../../../Data/Verification/regression/ds2/INLA_%s_uncorr_cells_fine'%name_dataset elif mesh_id == 2: dir_results = f'../../../../Data/Verification/regression/ds2/INLA_%s_uncorr_cells_medium'%name_dataset elif mesh_id == 3: dir_results = f'../../../../Data/Verification/regression/ds2/INLA_%s_uncorr_cells_coarse'%name_dataset #cell info fname_cellinfo = dir_syndata + '/' + 'CatalogNGAWest3CALite_cellinfo.csv' fname_distmat = dir_syndata + '/' + 'CatalogNGAWest3CALite_distancematrix.csv' #prefix for synthetic data and results prfx_syndata = 'CatalogNGAWest3CALite_synthetic' #regression results filename prefix prfx_results = f'%s_syndata'%name_dataset # dataset info ds_id = np.arange(1,6) # ++++++++++++++++++++++++++++++++++++++++ # USER NEEDS TO SPECIFY HYPERPARAMETERS OF SYNTHETIC DATASET # ++++++++++++++++++++++++++++++++++++++++ # hyper-parameters if corr_id == 1: # small correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, 'ell_1e':60, 'ell_1as':30, 'c_cap_erg': -0.011, 'omega_cap_mu': 0.005, 'omega_ca1p':0.004, 'omega_ca2p':0.002, 'ell_ca1p': 75, 'phi_0':0.4, 'tau_0':0.3 } elif corr_id == 2: #large correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, 'ell_1e':100, 'ell_1as':70, 'c_cap_erg': -0.02, 'omega_cap_mu': 0.008, 'omega_ca1p':0.005, 'omega_ca2p':0.003, 'ell_ca1p': 120, 'phi_0':0.4, 'tau_0':0.3} # ++++++++++++++++++++++++++++++++++++++++ # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #output filename sufix if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #kenel info if ker_id == 1: ker_suffix = '' elif ker_id == 2: ker_suffix = '_nexp' # ++++++++++++++++++++++++++++++++++++++++ #ploting options flag_report = True # Compare results # --------------------------- #load cell data df_cellinfo = pd.read_csv(fname_cellinfo).set_index('cellid') df_distmat = pd.read_csv(fname_distmat).set_index('rsn') #initialize misfit metrics dataframe df_misfit = pd.DataFrame(index=['Y%i'%d_id for d_id in ds_id]) #iterate over different datasets for d_id in ds_id: # Load Data #--- --- --- --- --- #file names #synthetic data fname_sdata_gmotion = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'data', synds_suffix, d_id) + '.csv' fname_sdata_atten = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'atten', synds_suffix, d_id) + '.csv' #regression results fname_reg_gmotion = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id, prfx_results, synds_suffix, d_id, 'residuals') + '.csv' fname_reg_coeff = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id, prfx_results, synds_suffix, d_id, 'coefficients') + '.csv' fname_reg_atten = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id, prfx_results, synds_suffix, d_id, 'catten') + '.csv' #load synthetic results df_sdata_gmotion = pd.read_csv(fname_sdata_gmotion).set_index('rsn') df_sdata_atten = pd.read_csv(fname_sdata_atten).set_index('cellid') #load regression results df_reg_gmotion = pd.read_csv(fname_reg_gmotion).set_index('rsn') df_reg_coeff = pd.read_csv(fname_reg_coeff).set_index('rsn') df_reg_atten = pd.read_csv(fname_reg_atten).set_index('cellid') # Processing #--- --- --- --- --- #keep only relevant columns from synthetic dataset df_sdata_gmotion = df_sdata_gmotion.reindex(df_reg_gmotion.index) df_sdata_atten = df_sdata_atten.reindex(df_reg_atten.index) #distance matrix for records of interest df_dmat = df_distmat.reindex(df_sdata_gmotion.index) #find unique earthqakes and stations eq_id, eq_idx, eq_nrec = np.unique(df_sdata_gmotion.eqid, return_index=True, return_counts=True) sta_id, sta_idx, sta_nrec = np.unique(df_sdata_gmotion.ssn, return_index=True, return_counts=True) #number of paths per cell cell_npath = np.sum(df_dmat.loc[:,df_reg_atten.cellname] > 0, axis=0) # Compute Root Mean Square Error #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_rms'] = CalcRMS(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_rms'] = CalcRMS(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_rms'] = CalcRMS(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_rms'] = CalcRMS(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_rms'] = CalcRMS(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Compute Divergence #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_KL'] = CalcLKDivergece(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_KL'] = CalcLKDivergece(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Output #--- --- --- --- --- #figure directory dir_fig = '%s%s/Y%i/figures_cmp/'%(dir_results, ker_suffix+synds_suffix, d_id) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #compare ground motion predictions #... ... ... ... ... ... #figure title fname_fig = 'Y%i_scatter_tot_res'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #median ax.scatter(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title('Comparison total residuals, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=35) ax.set_ylabel('Estimated', fontsize=35) ax.grid(which='both') ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-10,2]) ax.set_ylim([-10,2]) fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1e #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1e_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2,2]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['dc_1e_sig'].values[eq_idx], df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.5]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1as #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1as_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2,2]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1as_sig'].values[sta_idx], df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.5]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1bs #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1bs_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2,2]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1bs_sig'].values[sta_idx], df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_cap #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_cap_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-0.05,0.02]) ax.set_ylim([-0.05,0.02]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_atten['c_cap_sig'], df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.00,0.03]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_npath'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(cell_npath, df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of paths', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,5e4]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare Misfit Metrics # --------------------------- #summary directory dir_sum = '%s%s/summary/'%(dir_results,ker_suffix+synds_suffix) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #figure directory dir_fig = '%s/figures/'%(dir_sum) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save df_misfit.to_csv(dir_sum + 'misfit_summary.csv') #RMS misfit fname_fig = 'misfit_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_cap_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('RSME', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #KL divergence fname_fig = 'KLdiv_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_cap_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('KL divergence', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare hyper-paramters # --------------------------- #iterate over different datasets df_reg_hyp = list() df_reg_hyp_post = list() for d_id in ds_id: # Load Data #--- --- --- --- --- #regression hyperparamters results fname_reg_hyp = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id,prfx_results, synds_suffix, d_id, 'hyperparameters') + '.csv' fname_reg_hyp_post = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id,prfx_results, synds_suffix, d_id, 'hyperposterior') + '.csv' #load regression results df_reg_hyp.append( pd.read_csv(fname_reg_hyp, index_col=0) ) df_reg_hyp_post.append( pd.read_csv(fname_reg_hyp_post, index_col=0) ) # Omega_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 40 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1,E}$', fontsize=30) ax.set_xlabel('$\omega_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.25]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 30 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1a,S}$', fontsize=30) ax.set_xlabel('$\omega_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1bs #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1bs' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 60 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1b,S}$', fontsize=30) ax.set_xlabel('$\omega_{1b,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 0.02 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1,E}$', fontsize=30) ax.set_xlabel('$\ell_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 0.1 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1a,S}$', fontsize=30) ax.set_xlabel('$\ell_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Tau_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'tau_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 60 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\tau_{0}$', fontsize=30) ax.set_xlabel(r'$\tau_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Phi_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'phi_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 100 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\phi_{0}$', fontsize=30) ax.set_xlabel('$\phi_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.6]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_ca #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_cap' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 1500 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(np.sqrt(hyp['omega_ca1p']**2+hyp['omega_ca2p']**2), ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{ca,P}$', fontsize=30) ax.set_xlabel('$\omega_{ca,p}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.05]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # # Delta c_0 # #--- --- --- --- --- # #hyper-paramter name # name_hyp = 'dc_0' # #figure title # fname_fig = 'post_dist_' + name_hyp # #create figure # fig, ax = plt.subplots(figsize = (10,10)) # for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): # #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mean = 15 # #plot posterior dist # pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') # #plot true value # ymax_hyp = ymax_mean # # ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') # #edit figure # ax.set_title(r'Comparison $\delta c_{0}$', fontsize=30) # ax.set_xlabel('$\delta c_{0}$', fontsize=25) # ax.set_ylabel('probability density function ', fontsize=25) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=22) # ax.tick_params(axis='y', labelsize=22) # #plot limits # ax.set_xlim([-1,1]) # ax.set_ylim([0,ymax_hyp]) # #save figure # fig.tight_layout() # # fig.savefig( dir_fig + fname_fig + '.png' )
35,819
39.022346
160
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_uncorr_cells_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_uncorr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory # out_dir_sub = 'PYSTAN_NGAWest3CA_uncorr_cells' # out_dir_sub = 'PYSTAN_NGAWest3CA_uncorr_cells_chol' # out_dir_sub = 'PYSTAN_NGAWest3CA_uncorr_cells_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest3CA_uncorr_cells_chol_eff2' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,831
30.933333
103
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_cmdstan_model2_uncorr_cells_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model2_uncorr_cells_unbounded_hyp import RunStan from regression_cmdstan_model2_uncorr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient2.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_uncorr_cells' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_uncorr_cells_chol' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_uncorr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_uncorr_cells_chol_eff2' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_uncorr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,298
33.669355
109
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_cmdstan_model2_uncorr_cells_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model2_uncorr_cells_unbounded_hyp import RunStan from regression_cmdstan_model2_uncorr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient2.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CA_uncorr_cells' # out_dir_sub = 'CMDSTAN_NGAWest2CA_uncorr_cells_chol' # out_dir_sub = 'CMDSTAN_NGAWest2CA_uncorr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CA_uncorr_cells_chol_eff2' # out_dir_sub = 'CMDSTAN_NGAWest2CA_uncorr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,177
32.96748
109
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_cmdstan_model2_corr_cells_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model2_corr_cells_unbounded_hyp import RunStan from regression_cmdstan_model2_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient2.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest3CA_corr_cells' # out_dir_sub = 'CMDSTAN_NGAWest3CA_corr_cells_chol' # out_dir_sub = 'CMDSTAN_NGAWest3CA_corr_cells_chol_efficient' # out_dir_sub = 'CMDSTAN_NGAWest3CA_corr_cells_chol_efficient2' # out_dir_sub = 'CMDSTAN_NGAWest3CA_corr_cells_chol_efficient_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,078
32.710744
107
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_cmdstan_model2_corr_cells_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model2_corr_cells_unbounded_hyp import RunStan from regression_cmdstan_model2_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient2.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_corr_cells' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_corr_cells_chol' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff2' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,274
33.475806
107
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_uncorr_cells_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_uncorr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory #pystan 2 # out_dir_sub = 'PYSTAN_NGAWest2CA_uncorr_cells' # out_dir_sub = 'PYSTAN_NGAWest2CA_uncorr_cells_chol' # out_dir_sub = 'PYSTAN_NGAWest2CA_uncorr_cells_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest2CA_uncorr_cells_chol_eff2' #pystan 3 # out_dir_sub = 'PYSTAN3_NGAWest2CA_uncorr_cells' # out_dir_sub = 'PYSTAN3_NGAWest2CA_uncorr_cells_chol' # out_dir_sub = 'PYSTAN3_NGAWest2CA_uncorr_cells_chol_eff' # out_dir_sub = 'PYSTAN3_NGAWest2CA_uncorr_cells_chol_eff2' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,181
31.418605
103
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_cmdstan_model2_uncorr_cells_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model2_uncorr_cells_unbounded_hyp import RunStan from regression_cmdstan_model2_uncorr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_unbounded_hyp_chol_efficient2.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_uncorr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest3CA_uncorr_cells' # out_dir_sub = 'CMDSTAN_NGAWest3CA_uncorr_cells_chol' # out_dir_sub = 'CMDSTAN_NGAWest3CA_uncorr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest3CA_uncorr_cells_chol_eff2' # out_dir_sub = 'CMDSTAN_NGAWest3CA_uncorr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,084
32.760331
109
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_corr_cells_NGAWest3CA_sparse.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory out_dir_sub = 'PYSTAN_NGAWest3CA_corr_cells_chol_eff_sp' #stan parameters runstan_flag = True pystan_ver = 2 # pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,379
28.911504
105
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/comparison_stan_model2_uncorr_cells.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Aug 12 10:26:06 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load packages import os import sys import pathlib import glob import re #regular expression package import pickle #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick #user functions sys.path.insert(0,'../../../Python_lib/regression/') from pylib_stats import CalcRMS from pylib_stats import CalcLKDivergece # Define variables # --------------------------- # USER SETS DIRECTORIES AND FILE INFO OF SYNTHETIC DS AND REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #processed dataset # name_dataset = 'NGAWest2CANorth' name_dataset = 'NGAWest2CA' # name_dataset = 'NGAWest3CA' #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #package # 1: Pystan v2 # 2: Pystan v3 # 3: stancmd pkg_id = 1 #approximation type # 1: multivariate normal # 2: cholesky # 3: cholesky efficient # 4: cholesky efficient v2 # 5: cholesky efficient, sparse cells aprox_id = 3 #directories (synthetic dataset) if corr_id == 1: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds2_small_corr_len' elif corr_id == 2: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds2_large_corr_len' #cell info fname_cellinfo = dir_syndata + '/' + 'CatalogNGAWest3CALite_cellinfo.csv' fname_distmat = dir_syndata + '/' + 'CatalogNGAWest3CALite_distancematrix.csv' #directories (regression results) if pkg_id == 1: dir_results = f'../../../../Data/Verification/regression/ds2/PYSTAN_%s'%name_dataset elif pkg_id == 2: dir_results = f'../../../../Data/Verification/regression/ds2/PYSTAN3_%s'%name_dataset elif pkg_id == 3: dir_results = f'../../../../Data/Verification/regression/ds2/CMDSTAN_%s'%name_dataset # #directories (regression results) # if pkg_id == 1: # dir_results = f'../../../../Data/Verification/regression_old/ds2/PYSTAN_%s'%name_dataset # elif pkg_id == 2: # dir_results = f'../../../../Data/Verification/regression_old/ds2/PYSTAN3_%s'%name_dataset # elif pkg_id == 3: # dir_results = f'../../../../Data/Verification/regression_old/ds2/CMDSTAN_%s'%name_dataset #prefix for synthetic data and results prfx_syndata = 'CatalogNGAWest3CALite_synthetic' #regression results filename prefix prfx_results = f'%s_syndata'%name_dataset # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #output filename sufix (synthetic dataset) if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #output filename sufix (regression results) if aprox_id == 1: synds_suffix_stan = '_corr_cells' + synds_suffix elif aprox_id == 2: synds_suffix_stan = '_corr_cells' + '_chol' + synds_suffix elif aprox_id == 3: synds_suffix_stan = '_corr_cells' + '_chol_eff' + synds_suffix elif aprox_id == 4: synds_suffix_stan = '_corr_cells' + '_chol_eff2' + synds_suffix elif aprox_id == 5: synds_suffix_stan = '_corr_cells' + '_chol_eff_sp' + synds_suffix # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #output filename sufix (synthetic dataset) if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #output filename sufix (regression results) if aprox_id == 1: synds_suffix_stan = '_uncorr_cells' + synds_suffix elif aprox_id == 2: synds_suffix_stan = '_uncorr_cells' + '_chol' + synds_suffix elif aprox_id == 3: synds_suffix_stan = '_uncorr_cells' + '_chol_eff' + synds_suffix elif aprox_id == 4: synds_suffix_stan = '_uncorr_cells' + '_chol_eff2' + synds_suffix elif aprox_id == 5: synds_suffix_stan = '_uncorr_cells' + '_chol_eff_sp' + synds_suffix # dataset info ds_id = np.arange(1,6) # ++++++++++++++++++++++++++++++++++++++++ # USER NEEDS TO SPECIFY HYPERPARAMETERS OF SYNTHETIC DATASET # ++++++++++++++++++++++++++++++++++++++++ # hyper-parameters if corr_id == 1: # small correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, 'ell_1e':60, 'ell_1as':30, 'c_cap_erg': -0.011, 'omega_cap_mu': 0.005, 'omega_ca1p':0.004, 'omega_ca2p':0.002, 'ell_ca1p': 75, 'phi_0':0.4, 'tau_0':0.3 } elif corr_id == 2: #large correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, 'ell_1e':100, 'ell_1as':70, 'c_cap_erg': -0.02, 'omega_cap_mu': 0.008, 'omega_ca1p':0.005, 'omega_ca2p':0.003, 'ell_ca1p': 120, 'phi_0':0.4, 'tau_0':0.3} # ++++++++++++++++++++++++++++++++++++++++ #ploting options flag_report = True # Compare results # --------------------------- #load cell data df_cellinfo = pd.read_csv(fname_cellinfo).set_index('cellid') df_distmat = pd.read_csv(fname_distmat).set_index('rsn') #initialize misfit metrics dataframe df_misfit = pd.DataFrame(index=['Y%i'%d_id for d_id in ds_id]) #iterate over different datasets for d_id in ds_id: # Load Data #--- --- --- --- --- #file names #synthetic data fname_sdata_gmotion = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'data', synds_suffix, d_id) + '.csv' fname_sdata_atten = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'atten', synds_suffix, d_id) + '.csv' #regression results fname_reg_gmotion = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'residuals') + '.csv' fname_reg_coeff = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'coefficients') + '.csv' fname_reg_atten = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'catten') + '.csv' #load synthetic results df_sdata_gmotion = pd.read_csv(fname_sdata_gmotion).set_index('rsn') df_sdata_atten = pd.read_csv(fname_sdata_atten).set_index('cellid') #load regression results df_reg_gmotion = pd.read_csv(fname_reg_gmotion, index_col=0) df_reg_coeff = pd.read_csv(fname_reg_coeff, index_col=0) df_reg_atten = pd.read_csv(fname_reg_atten, index_col=0) # Processing #--- --- --- --- --- #keep only relevant columns from synthetic dataset df_sdata_gmotion = df_sdata_gmotion.reindex(df_reg_gmotion.index) df_sdata_atten = df_sdata_atten.reindex(df_reg_atten.index) #distance matrix for records of interest df_dmat = df_distmat.reindex(df_sdata_gmotion.index) #find unique earthqakes and stations eq_id, eq_idx, eq_nrec = np.unique(df_sdata_gmotion.eqid, return_index=True, return_counts=True) sta_id, sta_idx, sta_nrec = np.unique(df_sdata_gmotion.ssn, return_index=True, return_counts=True) #number of paths per cell cell_npath = np.sum(df_dmat.loc[:,df_reg_atten.cellname] > 0, axis=0) # Compute Root Mean Square Error #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_rms'] = CalcRMS(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_rms'] = CalcRMS(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_rms'] = CalcRMS(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_rms'] = CalcRMS(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_rms'] = CalcRMS(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Compute Divergence #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_KL'] = CalcLKDivergece(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_KL'] = CalcLKDivergece(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Output #--- --- --- --- --- #figure directory dir_fig = '%s%s/Y%i/figures_cmp/'%(dir_results,synds_suffix_stan,d_id) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #compare ground motion predictions #... ... ... ... ... ... #figure title fname_fig = 'Y%i_scatter_tot_res'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #median ax.scatter(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title('Comparison total residuals, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=35) ax.set_ylabel('Estimated', fontsize=35) ax.grid(which='both') ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-10,2]) ax.set_ylim([-10,2]) fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1e #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1e_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-.4,.4]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['dc_1e_sig'].values[eq_idx], df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1as #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1as_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1as_sig'].values[sta_idx], df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1bs #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1bs_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1bs_sig'].values[sta_idx], df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_cap #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_cap_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-0.05,0.02]) ax.set_ylim([-0.05,0.02]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_atten['c_cap_sig'], df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.00,0.03]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_npath'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(cell_npath, df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of paths', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,5e4]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare Misfit Metrics # --------------------------- #summary directory dir_sum = '%s%s/summary/'%(dir_results,synds_suffix_stan) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #figure directory dir_fig = '%s/figures/'%(dir_sum) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save df_misfit.to_csv(dir_sum + 'misfit_summary.csv') #RMS misfit fname_fig = 'misfit_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_cap_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('RSME', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #KL divergence fname_fig = 'KLdiv_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_cap_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('KL divergence', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare hyper-paramters # --------------------------- #iterate over different datasets df_reg_hyp = list() df_reg_hyp_post = list() for d_id in ds_id: # Load Data #--- --- --- --- --- #regression hyperparamters results fname_reg_hyp = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperparameters') + '.csv' fname_reg_hyp_post = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperposterior') + '.csv' #load regression results df_reg_hyp.append( pd.read_csv(fname_reg_hyp, index_col=0) ) df_reg_hyp_post.append( pd.read_csv(fname_reg_hyp_post, index_col=0) ) # Omega_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 40 ymax_mean = 40 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1,E}$', fontsize=30) ax.set_xlabel('$\omega_{1,E}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.25]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 30 ymax_mean = 30 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1a,S}$', fontsize=30) ax.set_xlabel('$\omega_{1a,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1bs #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1bs' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1b,S}$', fontsize=30) ax.set_xlabel('$\omega_{1b,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.02 ymax_mean = 0.02 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1,E}$', fontsize=30) ax.set_xlabel('$\ell_{1,E}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1a,S}$', fontsize=30) ax.set_xlabel('$\ell_{1a,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Tau_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'tau_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 150 ymax_mean = 150 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\tau_{0}$', fontsize=30) ax.set_xlabel(r'$\tau_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Phi_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'phi_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1000 ymax_mean = 1000 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\phi_{0}$', fontsize=30) ax.set_xlabel('$\phi_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.6]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_ca #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_cap' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1500 ymax_mean = 1500 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(np.sqrt(hyp['omega_ca1p']**2+hyp['omega_ca2p']**2), ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{ca,P}$', fontsize=30) ax.set_xlabel('$\omega_{ca,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.05]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # # Delta c_0 # #--- --- --- --- --- # #hyper-paramter name # name_hyp = 'dc_0' # #figure title # fname_fig = 'post_dist_' + name_hyp # #create figure # fig, ax = plt.subplots(figsize = (10,10)) # for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): # #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mean = 15 # #plot posterior dist # pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') # #plot true value # ymax_hyp = ymax_mean # # ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') # #edit figure # ax.set_title(r'Comparison $\delta c_{0}$', fontsize=30) # ax.set_xlabel('$\delta c_{0}$', fontsize=25) # ax.set_ylabel('probability density function ', fontsize=25) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=22) # ax.tick_params(axis='y', labelsize=22) # #plot limits # ax.set_xlim([-1,1]) # ax.set_ylim([0,ymax_hyp]) # #save figure # fig.tight_layout() # # fig.savefig( dir_fig + fname_fig + '.png' )
35,753
38.682575
153
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/comparison_model2_misfit.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 15 14:50:27 2022 @author: glavrent """ # Working directory and Packages # --------------------------- #load variables import os import sys import pathlib #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt #user functions def PlotRSMCmp(df_rms_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for k in df_rms_all: df_rms = df_rms_all[k] ds_id = np.array(range(len(df_rms))) ax.plot(ds_id, df_rms.loc[:,c_name+'_rms'], linestyle='-', marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('RMSE', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_rms.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax def PlotKLCmp(df_KL_all, c_name, fig_fname): #create figure axes fig, ax = plt.subplots(figsize = (10,10)) for k in df_KL_all: df_KL = df_KL_all[k] ds_id = np.array(range(len(df_KL))) ax.plot(ds_id, df_KL.loc[:,c_name+'_KL'], linestyle='-', marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim([0, max(0.50, max(ax.get_ylim()))]) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('KL divergence', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_KL.index) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='upper left', fontsize=32) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' ) return fig, ax # Define variables # --------------------------- # # Sparse Distance Matrix # # --- --- --- --- --- # # NGAWest 2 CA North # cmp_name = 'STAN_sparse_cmp_NGAWest2CA' # reg_title = ['STAN','STAN w/ sp dist matrix'] # reg_fname = ['CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len','CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_sp_small_corr_len'] # ylim_time = [0, 800] # NGAWest 2 CA cmp_name = 'STAN_sparse_cmp_NGAWest2CA' reg_title = ['STAN','STAN w/ sp dist matrix'] reg_fname = ['CMDSTAN_NGAWest2CA_corr_cells_chol_eff_small_corr_len','CMDSTAN_NGAWest2CA_corr_cells_chol_eff_sp_small_corr_len'] ylim_time = [0, 7000] # # Different Software # # --- --- --- --- --- # cmp_name = 'STAN_vs_INLA_cmp_NGAWest2CANorth' # reg_title = ['STAN corr. cells','STAN uncorr. cells','INLA uncorr. cells'] # reg_fname = ['CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len','CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len', # 'INLA_NGAWest2CANorth_uncorr_cells_coarse_small_corr_len'] # reg_fname = ['PYSTAN_NGAWest2CANorth_corr_cells_chol_eff_small_corr_len','PYSTAN_NGAWest2CANorth_uncorr_cells_chol_eff_small_corr_len', # 'INLA_NGAWest2CANorth_uncorr_cells_coarse_small_corr_len'] # ylim_time = [0, 800] #directories regressions # reg_dir = [f'../../../../Data/Verification/regression/ds2/%s/'%r_f for r_f in reg_fname] reg_dir = [f'../../../../Data/Verification/regression_old/ds2/%s/'%r_f for r_f in reg_fname] #directory output # dir_out = '../../../../Data/Verification/regression/ds2/comparisons/' dir_out = '../../../../Data/Verification/regression_old/ds2/comparisons/' # Load Data # --------------------------- #initialize misfit dataframe df_sum_misfit_all = {}; #read misfit info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename misfit info fname_sum = r_d + 'summary/misfit_summary.csv' #read KL score for coefficients df_sum_misfit_all[r_t] = pd.read_csv(fname_sum, index_col=0) #initialize run time dataframe df_runinfo_all = {}; #read run time info for k, (r_t, r_d) in enumerate(zip(reg_title, reg_dir)): #filename run time fname_runinfo = r_d + '/run_info.csv' #store calc time df_runinfo_all[r_t] = pd.read_csv(fname_runinfo) # Comparison Figures # --------------------------- pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # RMSE divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_RMSE'%(dir_out, cmp_name, c_name) #plotting PlotRSMCmp(df_sum_misfit_all , c_name, fig_fname); # KL divergence # --- --- --- --- --- #coefficient name c_name = 'nerg_tot' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1e' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1as' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); #coefficient name c_name = 'dc_1bs' #figure name fig_fname = '%s/%s_%s_KLdiv'%(dir_out, cmp_name, c_name) #plotting PlotKLCmp(df_sum_misfit_all , c_name, fig_fname); # Run Time # --- --- --- --- --- #run time figure fig_fname = '%s/%s_run_time'%(dir_out, cmp_name) #create figure axes fig, ax = plt.subplots(figsize = (10,10)) #iterate over different analyses for j, k in enumerate(df_runinfo_all): ds_id = df_runinfo_all[k].ds_id ds_name = ['Y%i'%d_i for d_i in ds_id] run_time = df_runinfo_all[k].run_time ax.plot(ds_id, run_time, marker='o', linewidth=2, markersize=10, label=k) #figure properties ax.set_ylim(ylim_time) ax.set_xlabel('synthetic dataset', fontsize=35) ax.set_ylabel('Run Time (min)', fontsize=35) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=ds_name) ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #legend ax.legend(loc='lower left', fontsize=32) # ax.legend(loc='upper left', fontsize=32) # ax.legend(loc='center left', bbox_to_anchor=(1, 0.5), fontsize=25) #save figure fig.tight_layout() fig.savefig( fig_fname + '.png' )
6,888
29.892377
140
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_cmdstan_model2_corr_cells_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model2_corr_cells_unbounded_hyp import RunStan from regression_cmdstan_model2_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Validation/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient2.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Validation/regression/ds2/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CA_corr_cells' # out_dir_sub = 'CMDSTAN_NGAWest2CA_corr_cells_chol' # out_dir_sub = 'CMDSTAN_NGAWest2CA_corr_cells_chol_efficient' # out_dir_sub = 'CMDSTAN_NGAWest2CA_corr_cells_chol_efficient2' # out_dir_sub = 'CMDSTAN_NGAWest2CA_corr_cells_chol_efficient_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
4,163
32.853659
107
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds2/main_pystan_model2_corr_cells_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model2_corr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds2' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model2_corr_cells_unbounded_hyp_chol_efficient2.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds2/' #output sub-directory # out_dir_sub = 'PYSTAN_NGAWest3CA_corr_cells' # out_dir_sub = 'PYSTAN_NGAWest3CA_corr_cells_chol' # out_dir_sub = 'PYSTAN_NGAWest3CA_corr_cells_chol_eff' # out_dir_sub = 'PYSTAN_NGAWest3CA_corr_cells_chol_eff2' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,811
30.766667
101
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_pystan_model3_uncorr_cells_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model3_uncorr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Validation/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Validation/regression/ds3/' #output sub-directory out_dir_sub = 'PYSTAN_NGAWest3CA_uncorr_cells_chol_eff' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,445
28.20339
100
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_cmdstan_model3_corr_cells_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model3_corr_cells_unbounded_hyp import RunStan # from regression_cmdstan_model3_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds3/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CA_corr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CA_corr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg= 0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,778
30.491667
107
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_cmdstan_model3_corr_cells_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model3_corr_cells_unbounded_hyp import RunStan # from regression_cmdstan_model3_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds3/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_corr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg= 0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,888
31.140496
107
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_pystan_model3_corr_cells_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model3_corr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Validation/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Validation/regression/ds3/' #output sub-directory out_dir_sub = 'PYSTAN_NGAWest3CA_corr_cells_chol_eff' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,441
28.169492
98
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_cmdstan_model3_uncorr_cells_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model3_uncorr_cells_unbounded_hyp import RunStan # from regression_cmdstan_model3_uncorr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds3/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_uncorr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CANorth_uncorr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg= 0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,899
31.231405
109
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/comparison_inla_model3_uncorr_cells.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Aug 12 10:26:06 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load packages import os import sys import pathlib import glob import re #regular expression package import pickle #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick #user functions sys.path.insert(0,'../../../Python_lib/regression/') from pylib_stats import CalcRMS from pylib_stats import CalcLKDivergece # Define variables # --------------------------- # USER SETS DIRECTORIES AND FILE INFO OF SYNTHETIC DS AND REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #processed dataset # name_dataset = 'NGAWest2CANorth' # name_dataset = 'NGAWest2CA' # name_dataset = 'NGAWest3CA' #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #kernel function # 1: Mattern kernel (alpha=2) # 2: Negative Exp (alpha=3/2) ker_id = 1 #mesh type # 1: Fine Mesh # 2: Medium Mesh # 3: Coarse Mesh mesh_id = 3 #directories (synthetic dataset) if corr_id == 1: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds3_small_corr_len' elif corr_id == 2: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds3_large_corr_len' #directories (regression results) if mesh_id == 1: dir_results = f'../../../../Data/Verification/regression/ds3/INLA_%s_uncorr_cells_fine'%name_dataset elif mesh_id == 2: dir_results = f'../../../../Data/Verification/regression/ds3/INLA_%s_uncorr_cells_medium'%name_dataset elif mesh_id == 3: dir_results = f'../../../../Data/Verification/regression/ds3/INLA_%s_uncorr_cells_coarse'%name_dataset #cell info fname_cellinfo = dir_syndata + '/' + 'CatalogNGAWest3CALite_cellinfo.csv' fname_distmat = dir_syndata + '/' + 'CatalogNGAWest3CALite_distancematrix.csv' #prefix for synthetic data and results prfx_syndata = 'CatalogNGAWest3CALite_synthetic' #regression results filename prefix prfx_results = f'%s_syndata'%name_dataset # dataset info ds_id = np.arange(1,6) # ++++++++++++++++++++++++++++++++++++++++ # USER NEEDS TO SPECIFY HYPERPARAMETERS OF SYNTHETIC DATASET # ++++++++++++++++++++++++++++++++++++++++ # hyper-parameters if corr_id == 1: # small correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, 'ell_1e':60, 'ell_1as':30, 'c_2_erg': -2.0, 'omega_2': 0.2, 'omega_2p': 0.15, 'ell_2p': 80, 'c_3_erg':-0.6, 'omega_3': 0.15, 'omega_3s': 0.15, 'ell_3s': 130, 'c_cap_erg': -0.011, 'omega_cap_mu': 0.005, 'omega_ca1p':0.004, 'omega_ca2p':0.002, 'ell_ca1p': 75, 'phi_0':0.3, 'tau_0':0.25 } elif corr_id == 2: # large correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, 'ell_1e':100, 'ell_1as':70, 'c_2_erg': -2.0, 'omega_2': 0.2, 'omega_2p': 0.15, 'ell_2e': 140, 'c_3_erg':-0.6, 'omega_3': 0.15, 'omega_3s': 0.15, 'ell_3s': 180, 'c_cap_erg': -0.02, 'omega_cap_mu': 0.008, 'omega_ca1p':0.005, 'omega_ca2p':0.003, 'ell_ca1p': 120, 'phi_0':0.3, 'tau_0':0.25} # ++++++++++++++++++++++++++++++++++++++++ # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #output filename sufix if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #kenel info if ker_id == 1: ker_suffix = '' elif ker_id == 2: ker_suffix = '_nexp' # ++++++++++++++++++++++++++++++++++++++++ #ploting options flag_report = True # Compare results # --------------------------- #load cell data df_cellinfo = pd.read_csv(fname_cellinfo).set_index('cellid') df_distmat = pd.read_csv(fname_distmat).set_index('rsn') #initialize misfit metrics dataframe df_misfit = pd.DataFrame(index=['Y%i'%d_id for d_id in ds_id]) #iterate over different datasets for d_id in ds_id: # Load Data #--- --- --- --- --- #file names #synthetic data fname_sdata_gmotion = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'data', synds_suffix, d_id) + '.csv' fname_sdata_atten = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'atten', synds_suffix, d_id) + '.csv' #regression results fname_reg_gmotion = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id, prfx_results, synds_suffix, d_id, 'residuals') + '.csv' fname_reg_coeff = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id, prfx_results, synds_suffix, d_id, 'coefficients') + '.csv' fname_reg_atten = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id, prfx_results, synds_suffix, d_id, 'catten') + '.csv' #load synthetic results df_sdata_gmotion = pd.read_csv(fname_sdata_gmotion).set_index('rsn') df_sdata_atten = pd.read_csv(fname_sdata_atten).set_index('cellid') #load regression results df_reg_gmotion = pd.read_csv(fname_reg_gmotion).set_index('rsn') df_reg_coeff = pd.read_csv(fname_reg_coeff).set_index('rsn') df_reg_atten = pd.read_csv(fname_reg_atten).set_index('cellid') # Processing #--- --- --- --- --- #keep only relevant columns from synthetic dataset df_sdata_gmotion = df_sdata_gmotion.reindex(df_reg_gmotion.index) df_sdata_atten = df_sdata_atten.reindex(df_reg_atten.index) #distance matrix for records of interest df_dmat = df_distmat.reindex(df_sdata_gmotion.index) #find unique earthqakes and stations eq_id, eq_idx, eq_nrec = np.unique(df_sdata_gmotion.eqid, return_index=True, return_counts=True) sta_id, sta_idx, sta_nrec = np.unique(df_sdata_gmotion.ssn, return_index=True, return_counts=True) #number of paths per cell cell_npath = np.sum(df_dmat.loc[:,df_reg_atten.cellname] > 0, axis=0) # Compute Root Mean Square Error #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_rms'] = CalcRMS(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_rms'] = CalcRMS(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_rms'] = CalcRMS(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_rms'] = CalcRMS(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_2p_rms'] = CalcRMS(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'c_3s_rms'] = CalcRMS(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_rms'] = CalcRMS(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Compute Divergence #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_KL'] = CalcLKDivergece(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_2p_KL'] = CalcLKDivergece(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'c_3s_KL'] = CalcLKDivergece(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_KL'] = CalcLKDivergece(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Output #--- --- --- --- --- #figure directory dir_fig = '%s%s/Y%i/figures_cmp/'%(dir_results, ker_suffix+synds_suffix, d_id) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #compare ground motion predictions #... ... ... ... ... ... #figure title fname_fig = 'Y%i_scatter_tot_res'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #median ax.scatter(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title('Comparison total residuals, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=35) ax.set_ylabel('Estimated', fontsize=35) ax.grid(which='both') ax.tick_params(axis='x', labelsize=32) ax.tick_params(axis='y', labelsize=32) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-10,2]) ax.set_ylim([-10,2]) fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1e #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1e_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2,2]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['dc_1e_sig'].values[eq_idx], df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.5]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1as #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1as_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2,2]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1as_sig'].values[sta_idx], df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.5]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1bs #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1bs_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2,2]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1bs_sig'].values[sta_idx], df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-2,2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_2p #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_2p_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2.3,-1.6]) ax.set_ylim([-2.3,-1.6]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_2p_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['c_2p_sig'].values[eq_idx], df_sdata_gmotion['c_2p'].values[eq_idx] - df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_2p_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['c_2p'].values[eq_idx] - df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_3s #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_3s_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.2,-.2]) ax.set_ylim([-1.2,-.2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_3s_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['c_3s_sig'].values[sta_idx], df_sdata_gmotion['c_3s'].values[sta_idx] - df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_3s_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(sta_nrec, df_sdata_gmotion['c_3s'].values[sta_idx] - df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_cap #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_cap_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-0.05,0.02]) ax.set_ylim([-0.05,0.02]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_atten['c_cap_sig'], df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.00,0.03]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_npath'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(cell_npath, df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of paths', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,5e4]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare Misfit Metrics # --------------------------- #summary directory dir_sum = '%s%s/summary/'%(dir_results,ker_suffix+synds_suffix) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #figure directory dir_fig = '%s/figures/'%(dir_sum) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save df_misfit.to_csv(dir_sum + 'misfit_summary.csv') #RMS misfit fname_fig = 'misfit_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_2p_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{2,E}$') ax.plot(ds_id, df_misfit.c_3s_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{3,S}$') ax.plot(ds_id, df_misfit.c_cap_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('RSME', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #KL divergence fname_fig = 'KLdiv_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_2p_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{2,P}$') ax.plot(ds_id, df_misfit.c_3s_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{3,S}$') ax.plot(ds_id, df_misfit.c_cap_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('KL divergence', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare hyper-paramters # --------------------------- #iterate over different datasets df_reg_hyp = list() df_reg_hyp_post = list() for d_id in ds_id: # Load Data #--- --- --- --- --- #regression hyperparamters results fname_reg_hyp = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id,prfx_results, synds_suffix, d_id, 'hyperparameters') + '.csv' fname_reg_hyp_post = '%s%s/Y%i/%s%s_Y%i_inla_%s'%(dir_results, ker_suffix+synds_suffix, d_id,prfx_results, synds_suffix, d_id, 'hyperposterior') + '.csv' #load regression results df_reg_hyp.append( pd.read_csv(fname_reg_hyp, index_col=0) ) df_reg_hyp_post.append( pd.read_csv(fname_reg_hyp_post, index_col=0) ) # Omega_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 40 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1,E}$', fontsize=30) ax.set_xlabel('$\omega_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.25]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 30 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1a,S}$', fontsize=30) ax.set_xlabel('$\omega_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1bs #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1bs' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 60 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1b,S}$', fontsize=30) ax.set_xlabel('$\omega_{1b,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_2p #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_2p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{2,P}$', fontsize=30) ax.set_xlabel('$\omega_{2,p}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_3s #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_3s' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{3,S}$', fontsize=30) ax.set_xlabel('$\omega_{3,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 0.02 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1,E}$', fontsize=30) ax.set_xlabel('$\ell_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 0.1 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1a,S}$', fontsize=30) ax.set_xlabel('$\ell_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_2p #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_2p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{2,P}$', fontsize=30) ax.set_xlabel('$\ell_{2,p}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_3s #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_3s' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{3,S}$', fontsize=30) ax.set_xlabel('$\ell_{3,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Tau_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'tau_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 60 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\tau_{0}$', fontsize=30) ax.set_xlabel(r'$\tau_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Phi_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'phi_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 100 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\phi_{0}$', fontsize=30) ax.set_xlabel('$\phi_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.6]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_ca #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_cap' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 ymax_mean = 1500 #plot posterior dist pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(np.sqrt(hyp['omega_ca1p']**2+hyp['omega_ca2p']**2), ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{ca,P}$', fontsize=30) ax.set_xlabel('$\omega_{ca,p}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.05]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # # Delta c_0 # #--- --- --- --- --- # #hyper-paramter name # name_hyp = 'dc_0' # #figure title # fname_fig = 'post_dist_' + name_hyp # #create figure # fig, ax = plt.subplots(figsize = (10,10)) # for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): # #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mean = 15 # #plot posterior dist # pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') # #plot true value # ymax_hyp = ymax_mean # # ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') # #edit figure # ax.set_title(r'Comparison $\delta c_{0}$', fontsize=30) # ax.set_xlabel('$\delta c_{0}$', fontsize=25) # ax.set_ylabel('probability density function ', fontsize=25) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=22) # ax.tick_params(axis='y', labelsize=22) # #plot limits # ax.set_xlim([-1,1]) # ax.set_ylim([0,ymax_hyp]) # #save figure # fig.tight_layout() # # fig.savefig( dir_fig + fname_fig + '.png' )
47,785
39.088926
160
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_pystan_model3_uncorr_cells_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model3_uncorr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Validation/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Validation/regression/ds3/' #output sub-directory out_dir_sub = 'PYSTAN_NGAWest2CA_uncorr_cells_chol_eff' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,534
28.458333
100
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_pystan_model3_corr_cells_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model3_corr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Validation/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Validation/regression/ds3/' #output sub-directory out_dir_sub = 'PYSTAN_NGAWest2CA_corr_cells_chol_eff' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,530
28.425
98
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/comparison_stan_model3_corr_cells.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Aug 12 10:26:06 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load packages import os import sys import pathlib import glob import re #regular expression package import pickle #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick #user functions sys.path.insert(0,'../../../Python_lib/regression/') from pylib_stats import CalcRMS from pylib_stats import CalcLKDivergece # Define variables # --------------------------- # USER SETS DIRECTORIES AND FILE INFO OF SYNTHETIC DS AND REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #processed dataset # name_dataset = 'NGAWest2CANorth' # name_dataset = 'NGAWest2CA' # name_dataset = 'NGAWest3CA' #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #package # 1: Pystan v2 # 2: Pystan v3 # 3: stancmd pkg_id = 3 #approximation type # 1: multivariate normal # 2: cholesky # 3: cholesky efficient # 4: cholesky efficient v2 # 5: cholesky efficient, sparse cells aprox_id = 3 #directories (synthetic dataset) if corr_id == 1: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds3_small_corr_len' elif corr_id == 2: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds3_large_corr_len' #cell info fname_cellinfo = dir_syndata + '/' + 'CatalogNGAWest3CALite_cellinfo.csv' fname_distmat = dir_syndata + '/' + 'CatalogNGAWest3CALite_distancematrix.csv' #directories (regression results) if pkg_id == 1: dir_results = f'../../../../Data/Verification/regression/ds3/PYSTAN_%s'%name_dataset elif pkg_id == 2: dir_results = f'../../../../Data/Verification/regression/ds3/PYSTAN3_%s'%name_dataset elif pkg_id == 3: dir_results = f'../../../../Data/Verification/regression/ds3/CMDSTAN_%s'%name_dataset #directories (regression results - old results) if pkg_id == 1: dir_results = f'../../../../Data/Verification/regression_old/ds3/PYSTAN_%s'%name_dataset elif pkg_id == 2: dir_results = f'../../../../Data/Verification/regression_old/ds3/PYSTAN3_%s'%name_dataset elif pkg_id == 3: dir_results = f'../../../../Data/Verification/regression_old/ds3/CMDSTAN_%s'%name_dataset #prefix for synthetic data and results prfx_syndata = 'CatalogNGAWest3CALite_synthetic' #regression results filename prefix prfx_results = f'%s_syndata'%name_dataset # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #output filename sufix (synthetic dataset) if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #output filename sufix (regression results) if aprox_id == 1: synds_suffix_stan = '_corr_cells' + synds_suffix elif aprox_id == 2: synds_suffix_stan = '_corr_cells' + '_chol' + synds_suffix elif aprox_id == 3: synds_suffix_stan = '_corr_cells' + '_chol_eff' + synds_suffix elif aprox_id == 4: synds_suffix_stan = '_corr_cells' + '_chol_eff2' + synds_suffix elif aprox_id == 5: synds_suffix_stan = '_corr_cells' + '_chol_eff_sp' + synds_suffix # dataset info # ds_id = np.arange(1,6) ds_id = np.arange(1,2) # ++++++++++++++++++++++++++++++++++++++++ # USER NEEDS TO SPECIFY HYPERPARAMETERS OF SYNTHETIC DATASET # ++++++++++++++++++++++++++++++++++++++++ # hyper-parameters if corr_id == 1: # small correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, 'ell_1e':60, 'ell_1as':30, 'c_2_erg': -2.0, 'omega_2': 0.2, 'omega_2p': 0.15, 'ell_2p': 80, 'c_3_erg':-0.6, 'omega_3': 0.15, 'omega_3s': 0.15, 'ell_3s': 130, 'c_cap_erg': -0.011, 'omega_cap_mu': 0.005, 'omega_ca1p':0.004, 'omega_ca2p':0.002, 'ell_ca1p': 75, 'phi_0':0.3, 'tau_0':0.25 } elif corr_id == 2: # large correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, 'ell_1e':100, 'ell_1as':70, 'c_2_erg': -2.0, 'omega_2': 0.2, 'omega_2p': 0.15, 'ell_2e': 140, 'c_3_erg':-0.6, 'omega_3': 0.15, 'omega_3s': 0.15, 'ell_3s': 180, 'c_cap_erg': -0.02, 'omega_cap_mu': 0.008, 'omega_ca1p':0.005, 'omega_ca2p':0.003, 'ell_ca1p': 120, 'phi_0':0.3, 'tau_0':0.25} # ++++++++++++++++++++++++++++++++++++++++ #ploting options flag_report = True # Compare results # --------------------------- #load cell data df_cellinfo = pd.read_csv(fname_cellinfo).set_index('cellid') df_distmat = pd.read_csv(fname_distmat).set_index('rsn') #initialize misfit metrics dataframe df_misfit = pd.DataFrame(index=['Y%i'%d_id for d_id in ds_id]) #iterate over different datasets for d_id in ds_id: # Load Data #--- --- --- --- --- #file names #synthetic data fname_sdata_gmotion = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'data', synds_suffix, d_id) + '.csv' fname_sdata_atten = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'atten', synds_suffix, d_id) + '.csv' #regression results fname_reg_gmotion = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'residuals') + '.csv' fname_reg_coeff = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'coefficients') + '.csv' fname_reg_atten = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'catten') + '.csv' #load synthetic results df_sdata_gmotion = pd.read_csv(fname_sdata_gmotion).set_index('rsn') df_sdata_atten = pd.read_csv(fname_sdata_atten).set_index('cellid') #load regression results df_reg_gmotion = pd.read_csv(fname_reg_gmotion, index_col=0) df_reg_coeff = pd.read_csv(fname_reg_coeff, index_col=0) df_reg_atten = pd.read_csv(fname_reg_atten, index_col=0) # Processing #--- --- --- --- --- #keep only relevant columns from synthetic dataset df_sdata_gmotion = df_sdata_gmotion.reindex(df_reg_gmotion.index) df_sdata_atten = df_sdata_atten.reindex(df_reg_atten.index) #distance matrix for records of interest df_dmat = df_distmat.reindex(df_sdata_gmotion.index) #find unique earthqakes and stations eq_id, eq_idx, eq_nrec = np.unique(df_sdata_gmotion.eqid, return_index=True, return_counts=True) sta_id, sta_idx, sta_nrec = np.unique(df_sdata_gmotion.ssn, return_index=True, return_counts=True) #number of paths per cell cell_npath = np.sum(df_dmat.loc[:,df_reg_atten.cellname] > 0, axis=0) # Compute Root Mean Square Error #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_rms'] = CalcRMS(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_rms'] = CalcRMS(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_rms'] = CalcRMS(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_rms'] = CalcRMS(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_2p_rms'] = CalcRMS(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'c_3s_rms'] = CalcRMS(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_rms'] = CalcRMS(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Compute Divergence #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_KL'] = CalcLKDivergece(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_2p_KL'] = CalcLKDivergece(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'c_3s_KL'] = CalcLKDivergece(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_KL'] = CalcLKDivergece(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Output #--- --- --- --- --- #figure directory dir_fig = '%s%s/Y%i/figures_cmp/'%(dir_results,synds_suffix_stan,d_id) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #compare ground motion predictions #... ... ... ... ... ... #figure title fname_fig = 'Y%i_scatter_tot_res'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #median ax.scatter(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title('Comparison total residuals, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-10,2]) ax.set_ylim([-10,2]) fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1e #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1e_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-.4,.4]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['dc_1e_sig'].values[eq_idx], df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1as #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1as_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1as_sig'].values[sta_idx], df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1bs #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1bs_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1bs_sig'].values[sta_idx], df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_2p #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_2p_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2.3,-1.6]) ax.set_ylim([-2.3,-1.6]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_2p_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['c_2p_sig'].values[eq_idx], df_sdata_gmotion['c_2p'].values[eq_idx] - df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_2p_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['c_2p'].values[eq_idx] - df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_3s #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_3s_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.2,-.2]) ax.set_ylim([-1.2,-.2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_3s_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['c_3s_sig'].values[sta_idx], df_sdata_gmotion['c_3s'].values[sta_idx] - df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_3s_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(sta_nrec, df_sdata_gmotion['c_3s'].values[sta_idx] - df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_cap #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_cap_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-0.05,0.02]) ax.set_ylim([-0.05,0.02]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_atten['c_cap_sig'], df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.00,0.03]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_npath'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(cell_npath, df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of paths', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,5e4]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare Misfit Metrics # --------------------------- #summary directory dir_sum = '%s%s/summary/'%(dir_results,synds_suffix_stan) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #figure directory dir_fig = '%s/figures/'%(dir_sum) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save df_misfit.to_csv(dir_sum + 'misfit_summary.csv') #RMS misfit fname_fig = 'misfit_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_2p_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{2,E}$') ax.plot(ds_id, df_misfit.c_3s_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{3,S}$') ax.plot(ds_id, df_misfit.c_cap_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('RSME', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #KL divergence fname_fig = 'KLdiv_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_2p_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{2,P}$') ax.plot(ds_id, df_misfit.c_3s_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{3,S}$') ax.plot(ds_id, df_misfit.c_cap_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('KL divergence', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare hyper-paramters # --------------------------- #iterate over different datasets df_reg_hyp = list() df_reg_hyp_post = list() for d_id in ds_id: # Load Data #--- --- --- --- --- #regression hyperparamters results fname_reg_hyp = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperparameters') + '.csv' fname_reg_hyp_post = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperposterior') + '.csv' #load regression results df_reg_hyp.append( pd.read_csv(fname_reg_hyp, index_col=0) ) df_reg_hyp_post.append( pd.read_csv(fname_reg_hyp_post, index_col=0) ) #figure directory dir_fig = '%s%s/figures_cmp_hyp/'%(dir_results,synds_suffix_stan) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) # Omega_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 40 ymax_mean = 40 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1,E}$', fontsize=30) ax.set_xlabel('$\omega_{1,E}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.25]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 30 ymax_mean = 30 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1a,S}$', fontsize=30) ax.set_xlabel('$\omega_{1a,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1bs #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1bs' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1b,S}$', fontsize=30) ax.set_xlabel('$\omega_{1b,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_2p #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_2p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{2,P}$', fontsize=30) ax.set_xlabel('$\omega_{2,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_3s #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_3s' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{3,S}$', fontsize=30) ax.set_xlabel('$\omega_{3,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.02 ymax_mean = 0.02 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1,E}$', fontsize=30) ax.set_xlabel('$\ell_{1,E}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1a,S}$', fontsize=30) ax.set_xlabel('$\ell_{1a,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_2p #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_2p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{2,P}$', fontsize=30) ax.set_xlabel('$\ell_{2,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_3s #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_3s' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{3,S}$', fontsize=30) ax.set_xlabel('$\ell_{3,S}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Tau_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'tau_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 150 ymax_mean = 150 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\tau_{0}$', fontsize=30) ax.set_xlabel(r'$\tau_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Phi_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'phi_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1000 ymax_mean = 1000 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\phi_{0}$', fontsize=30) ax.set_xlabel('$\phi_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.6]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_ca1p #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_ca1p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.02 ymax_mean = 0.02 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{ca1,P}$', fontsize=30) ax.set_xlabel('$\ell_{ca1,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_ca1 #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_ca1p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1500 ymax_mean = 1500 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp['omega_ca2p'], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{ca1,P}$', fontsize=30) ax.set_xlabel('$\omega_{ca1,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.05]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_ca2 #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_ca2p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1500 ymax_mean = 1500 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp['omega_ca2p'], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{ca2,p}$', fontsize=30) ax.set_xlabel('$\omega_{ca2,P}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.05]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # # Delta c_0 # #--- --- --- --- --- # #hyper-paramter name # name_hyp = 'dc_0' # #figure title # fname_fig = 'post_dist_' + name_hyp # #create figure # fig, ax = plt.subplots(figsize = (10,10)) # for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): # #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mean = 15 # #plot posterior dist # pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') # #plot true value # ymax_hyp = ymax_mean # # ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') # #edit figure # ax.set_title(r'Comparison $\delta c_{0}$', fontsize=30) # ax.set_xlabel('$\delta c_{0}$', fontsize=25) # ax.set_ylabel('probability density function ', fontsize=25) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=22) # ax.tick_params(axis='y', labelsize=22) # #plot limits # ax.set_xlim([-1,1]) # ax.set_ylim([0,ymax_hyp]) # #save figure # fig.tight_layout() # # fig.savefig( dir_fig + fname_fig + '.png' )
49,612
38.658673
153
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_cmdstan_model3_uncorr_cells_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model3_uncorr_cells_unbounded_hyp import RunStan # from regression_cmdstan_model3_uncorr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_unbounded_hyp_chol_efficient.stan' sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds3/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest3CA_uncorr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest3CA_uncorr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg= 0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,698
30.347458
107
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_cmdstan_model3_uncorr_cells_NGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model3_uncorr_cells_unbounded_hyp import RunStan # from regression_cmdstan_model3_uncorr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds3/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest2CA_uncorr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest2CA_uncorr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg= 0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,789
30.583333
109
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_cmdstan_model3_corr_cells_NGAWest3CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Dec 29 15:16:15 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/cmdstan/') # from regression_cmdstan_model3_corr_cells_unbounded_hyp import RunStan # from regression_cmdstan_model3_corr_cells_sparse_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Verification/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model # sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_unbounded_hyp_chol_efficient.stan' # sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_sparse_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest3CA_syndata' #main output directory out_dir_main = '../../../../Data/Verification/regression/ds3/' #output sub-directory # out_dir_sub = 'CMDSTAN_NGAWest3CA_corr_cells_chol_eff' # out_dir_sub = 'CMDSTAN_NGAWest3CA_corr_cells_chol_eff_sp' #stan parameters res_name = 'tot' n_iter_warmup = 500 n_iter_sampling = 500 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg= 0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only NGAWest2 records df_flatfile = df_flatfile.loc[df_flatfile.dsid==0,:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, n_iter_warmup=n_iter_warmup, n_iter_sampling=n_iter_sampling, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, stan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,777
30.747899
107
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/comparison_stan_model3_uncorr_cells.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Aug 12 10:26:06 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load packages import os import sys import pathlib import glob import re #regular expression package import pickle #arithmetic libraries import numpy as np #statistics libraries import pandas as pd #plot libraries import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.ticker import AutoLocator as plt_autotick #user functions sys.path.insert(0,'../../../Python_lib/regression/') from pylib_stats import CalcRMS from pylib_stats import CalcLKDivergece # Define variables # --------------------------- # USER SETS DIRECTORIES AND FILE INFO OF SYNTHETIC DS AND REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #processed dataset # name_dataset = 'NGAWest2CANorth' # name_dataset = 'NGAWest2CA' # name_dataset = 'NGAWest3CA' #correlation info # 1: Small Correlation Lengths # 2: Large Correlation Lenghts corr_id = 1 #package # 1: Pystan v2 # 2: Pystan v3 # 3: stancmd pkg_id = 3 #approximation type # 1: multivariate normal # 2: cholesky # 3: cholesky efficient # 4: cholesky efficient v2 # 5: cholesky efficient, sparse cells aprox_id = 3 #directories (synthetic dataset) if corr_id == 1: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds3_small_corr_len' elif corr_id == 2: dir_syndata = '../../../../Data/Verification/synthetic_datasets/ds3_large_corr_len' #cell info fname_cellinfo = dir_syndata + '/' + 'CatalogNGAWest3CALite_cellinfo.csv' fname_distmat = dir_syndata + '/' + 'CatalogNGAWest3CALite_distancematrix.csv' #directories (regression results) if pkg_id == 1: dir_results = f'../../../../Data/Verification/regression/ds3/PYSTAN_%s'%name_dataset elif pkg_id == 2: dir_results = f'../../../../Data/Verification/regression/ds3/PYSTAN3_%s'%name_dataset elif pkg_id == 3: dir_results = f'../../../../Data/Verification/regression/ds3/CMDSTAN_%s'%name_dataset #directories (regression results - old results) if pkg_id == 1: dir_results = f'../../../../Data/Verification/regression_old/ds3/PYSTAN_%s'%name_dataset elif pkg_id == 2: dir_results = f'../../../../Data/Verification/regression_old/ds3/PYSTAN3_%s'%name_dataset elif pkg_id == 3: dir_results = f'../../../../Data/Verification/regression_old/ds3/CMDSTAN_%s'%name_dataset #prefix for synthetic data and results prfx_syndata = 'CatalogNGAWest3CALite_synthetic' #regression results filename prefix prfx_results = f'%s_syndata'%name_dataset # FILE INFO FOR REGRESSION RESULTS # ++++++++++++++++++++++++++++++++++++++++ #output filename sufix (synthetic dataset) if corr_id == 1: synds_suffix = '_small_corr_len' elif corr_id == 2: synds_suffix = '_large_corr_len' #output filename sufix (regression results) if aprox_id == 1: synds_suffix_stan = '_uncorr_cells' + synds_suffix elif aprox_id == 2: synds_suffix_stan = '_uncorr_cells' + '_chol' + synds_suffix elif aprox_id == 3: synds_suffix_stan = '_uncorr_cells' + '_chol_eff' + synds_suffix elif aprox_id == 4: synds_suffix_stan = '_uncorr_cells' + '_chol_eff2' + synds_suffix elif aprox_id == 5: synds_suffix_stan = '_uncorr_cells' + '_chol_eff_sp' + synds_suffix # dataset info ds_id = np.arange(1,2) # ++++++++++++++++++++++++++++++++++++++++ # USER NEEDS TO SPECIFY HYPERPARAMETERS OF SYNTHETIC DATASET # ++++++++++++++++++++++++++++++++++++++++ # hyper-parameters if corr_id == 1: # small correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.1, 'omega_1as': 0.35, 'omega_1bs': 0.25, 'ell_1e':60, 'ell_1as':30, 'c_2_erg': -2.0, 'omega_2': 0.2, 'omega_2p': 0.15, 'ell_2p': 80, 'c_3_erg':-0.6, 'omega_3': 0.15, 'omega_3s': 0.15, 'ell_3s': 130, 'c_cap_erg': -0.011, 'omega_cap_mu': 0.005, 'omega_ca1p':0.004, 'omega_ca2p':0.002, 'ell_ca1p': 75, 'phi_0':0.3, 'tau_0':0.25 } elif corr_id == 2: # large correlation lengths hyp = {'omega_0': 0.1, 'omega_1e':0.2, 'omega_1as': 0.4, 'omega_1bs': 0.3, 'ell_1e':100, 'ell_1as':70, 'c_2_erg': -2.0, 'omega_2': 0.2, 'omega_2p': 0.15, 'ell_2e': 140, 'c_3_erg':-0.6, 'omega_3': 0.15, 'omega_3s': 0.15, 'ell_3s': 180, 'c_cap_erg': -0.02, 'omega_cap_mu': 0.008, 'omega_ca1p':0.005, 'omega_ca2p':0.003, 'ell_ca1p': 120, 'phi_0':0.3, 'tau_0':0.25} # ++++++++++++++++++++++++++++++++++++++++ #ploting options flag_report = True # Compare results # --------------------------- #load cell data df_cellinfo = pd.read_csv(fname_cellinfo).set_index('cellid') df_distmat = pd.read_csv(fname_distmat).set_index('rsn') #initialize misfit metrics dataframe df_misfit = pd.DataFrame(index=['Y%i'%d_id for d_id in ds_id]) #iterate over different datasets for d_id in ds_id: # Load Data #--- --- --- --- --- #file names #synthetic data fname_sdata_gmotion = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'data', synds_suffix, d_id) + '.csv' fname_sdata_atten = '%s/%s_%s%s_Y%i'%(dir_syndata, prfx_syndata, 'atten', synds_suffix, d_id) + '.csv' #regression results fname_reg_gmotion = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'residuals') + '.csv' fname_reg_coeff = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'coefficients') + '.csv' fname_reg_atten = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results, synds_suffix_stan, d_id, prfx_results, synds_suffix, d_id, 'catten') + '.csv' #load synthetic results df_sdata_gmotion = pd.read_csv(fname_sdata_gmotion).set_index('rsn') df_sdata_atten = pd.read_csv(fname_sdata_atten).set_index('cellid') #load regression results df_reg_gmotion = pd.read_csv(fname_reg_gmotion, index_col=0) df_reg_coeff = pd.read_csv(fname_reg_coeff, index_col=0) df_reg_atten = pd.read_csv(fname_reg_atten, index_col=0) # Processing #--- --- --- --- --- #keep only relevant columns from synthetic dataset df_sdata_gmotion = df_sdata_gmotion.reindex(df_reg_gmotion.index) df_sdata_atten = df_sdata_atten.reindex(df_reg_atten.index) #distance matrix for records of interest df_dmat = df_distmat.reindex(df_sdata_gmotion.index) #find unique earthqakes and stations eq_id, eq_idx, eq_nrec = np.unique(df_sdata_gmotion.eqid, return_index=True, return_counts=True) sta_id, sta_idx, sta_nrec = np.unique(df_sdata_gmotion.ssn, return_index=True, return_counts=True) #number of paths per cell cell_npath = np.sum(df_dmat.loc[:,df_reg_atten.cellname] > 0, axis=0) # Compute Root Mean Square Error #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_rms'] = CalcRMS(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_rms'] = CalcRMS(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_rms'] = CalcRMS(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_rms'] = CalcRMS(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_2p_rms'] = CalcRMS(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'c_3s_rms'] = CalcRMS(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_rms'] = CalcRMS(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Compute Divergence #--- --- --- --- --- df_misfit.loc['Y%i'%d_id,'nerg_tot_KL'] = CalcLKDivergece(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) df_misfit.loc['Y%i'%d_id,'dc_1e_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'dc_1as_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'dc_1bs_KL'] = CalcLKDivergece(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_2p_KL'] = CalcLKDivergece(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) df_misfit.loc['Y%i'%d_id,'c_3s_KL'] = CalcLKDivergece(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) df_misfit.loc['Y%i'%d_id,'c_cap_KL'] = CalcLKDivergece(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) # Output #--- --- --- --- --- #figure directory dir_fig = '%s%s/Y%i/figures_cmp/'%(dir_results,synds_suffix_stan,d_id) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #compare ground motion predictions #... ... ... ... ... ... #figure title fname_fig = 'Y%i_scatter_tot_res'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #median ax.scatter(df_sdata_gmotion.nerg_gm.values, df_reg_gmotion.nerg_mu.values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title('Comparison total residuals, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-10,2]) ax.set_ylim([-10,2]) fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1e #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1e_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1e'].values[eq_idx], df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-.4,.4]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['dc_1e_sig'].values[eq_idx], df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1e_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['dc_1e'].values[eq_idx] - df_reg_coeff['dc_1e_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1,E}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1as #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1as_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1as'].values[sta_idx], df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1as_sig'].values[sta_idx], df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1as_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1as'].values[sta_idx] - df_reg_coeff['dc_1as_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1a,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare dc_1bs #... ... ... ... ... ... #figure title fname_fig = 'Y%i_dc_1bs_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['dc_1bs'].values[sta_idx], df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.5,1.5]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(df_reg_coeff['dc_1bs_sig'].values[sta_idx], df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.4]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_dc_1bs_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #accuray ax.scatter(sta_nrec, df_sdata_gmotion['dc_1bs'].values[sta_idx] - df_reg_coeff['dc_1bs_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $\delta c_{1b,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,1000]) ax.set_ylim([-1.5,1.5]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_2p #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_2p_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['c_2p'].values[eq_idx], df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-2.3,-1.6]) ax.set_ylim([-2.3,-1.6]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_2p_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['c_2p_sig'].values[eq_idx], df_sdata_gmotion['c_2p'].values[eq_idx] - df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.15]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_2p_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(eq_nrec, df_sdata_gmotion['c_2p'].values[eq_idx] - df_reg_coeff['c_2p_mean'].values[eq_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{2,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_3s #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_3s_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_gmotion['c_3s'].values[sta_idx], df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-1.2,-.2]) ax.set_ylim([-1.2,-.2]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_3s_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_coeff['c_3s_sig'].values[sta_idx], df_sdata_gmotion['c_3s'].values[sta_idx] - df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0,.3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_3s_nrec'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(sta_nrec, df_sdata_gmotion['c_3s'].values[sta_idx] - df_reg_coeff['c_3s_mean'].values[sta_idx]) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{3,S}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of records', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.9,1e3]) ax.set_ylim([-.4,.4]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #compare c_cap #... ... ... ... ... ... #figure title fname_fig = 'Y%i_c_cap_scatter'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_sdata_atten['c_cap'].values, df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=1, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Synthetic dataset', fontsize=25) ax.set_ylabel('Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # plt_lim = np.array([ax.get_xlim(), ax.get_ylim()]) # plt_lim = (plt_lim[:,0].min(), plt_lim[:,1].max()) # ax.set_xlim(plt_lim) # ax.set_ylim(plt_lim) ax.set_xlim([-0.05,0.02]) ax.set_ylim([-0.05,0.02]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_accuracy'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(df_reg_atten['c_cap_sig'], df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Standard Deviation', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([0.00,0.03]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #figure title fname_fig = 'Y%i_c_cap_npath'%d_id #create figure fig, ax = plt.subplots(figsize = (10,10)) #coefficient scatter ax.scatter(cell_npath, df_sdata_atten['c_cap'].values - df_reg_atten['c_cap_mean'].values) ax.axline((0,0), slope=0, color="black", linestyle="--") #edit figure if not flag_report: ax.set_title(r'Comparison $c_{ca,P}$, Y: %i'%d_id, fontsize=30) ax.set_xlabel('Number of paths', fontsize=25) ax.set_ylabel('Actual - Estimated', fontsize=25) ax.grid(which='both') ax.set_xscale('log') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits # ax.set_ylim(np.abs(ax.get_ylim()).max()*np.array([-1,1])) ax.set_xlim([.9,5e4]) ax.set_ylim([-0.04,0.04]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare Misfit Metrics # --------------------------- #summary directory dir_sum = '%s%s/summary/'%(dir_results,synds_suffix_stan) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #figure directory dir_fig = '%s/figures/'%(dir_sum) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save df_misfit.to_csv(dir_sum + 'misfit_summary.csv') #RMS misfit fname_fig = 'misfit_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_2p_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{2,E}$') ax.plot(ds_id, df_misfit.c_3s_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{3,S}$') ax.plot(ds_id, df_misfit.c_cap_rms, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('RSME', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) #KL divergence fname_fig = 'KLdiv_score' #plot KL divergence fig, ax = plt.subplots(figsize = (10,10)) ax.plot(ds_id, df_misfit.nerg_tot_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label= 'tot nerg') ax.plot(ds_id, df_misfit.dc_1e_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1,E}$') ax.plot(ds_id, df_misfit.dc_1as_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1a,S}$') ax.plot(ds_id, df_misfit.dc_1bs_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$\delta c_{1b,S}$') ax.plot(ds_id, df_misfit.c_2p_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{2,P}$') ax.plot(ds_id, df_misfit.c_3s_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{3,S}$') ax.plot(ds_id, df_misfit.c_cap_KL, linestyle='-', marker='o', linewidth=2, markersize=10, label=r'$c_{ca,P}$') #figure properties ax.set_ylim([0,0.50]) ax.set_xlabel('synthetic dataset', fontsize=25) ax.set_ylabel('KL divergence', fontsize=25) ax.grid(which='both') ax.set_xticks(ds_id) ax.set_xticklabels(labels=df_misfit.index) ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #legend ax.legend(loc='upper left', fontsize=25) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Compare hyper-paramters # --------------------------- #iterate over different datasets df_reg_hyp = list() df_reg_hyp_post = list() for d_id in ds_id: # Load Data #--- --- --- --- --- #regression hyperparamters results fname_reg_hyp = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperparameters') + '.csv' fname_reg_hyp_post = '%s%s/Y%i/%s%s_Y%i_stan_%s'%(dir_results,synds_suffix_stan, d_id,prfx_results, synds_suffix, d_id, 'hyperposterior') + '.csv' #load regression results df_reg_hyp.append( pd.read_csv(fname_reg_hyp, index_col=0) ) df_reg_hyp_post.append( pd.read_csv(fname_reg_hyp_post, index_col=0) ) #figure directory dir_fig = '%s%s/figures_cmp_hyp/'%(dir_results,synds_suffix_stan) pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) # Omega_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 40 ymax_mean = 40 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1,E}$', fontsize=30) ax.set_xlabel('$\omega_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.25]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 30 ymax_mean = 30 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1a,S}$', fontsize=30) ax.set_xlabel('$\omega_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_1bs #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_1bs' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{1b,S}$', fontsize=30) ax.set_xlabel('$\omega_{1b,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_2p #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_2p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{2,P}$', fontsize=30) ax.set_xlabel('$\omega_{2,p}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_3s #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_3s' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 60 ymax_mean = 60 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{3,S}$', fontsize=30) ax.set_xlabel('$\omega_{3,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1e #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1e' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.02 ymax_mean = 0.02 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1,E}$', fontsize=30) ax.set_xlabel('$\ell_{1,e}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,500]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_1as #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_1as' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{1a,S}$', fontsize=30) ax.set_xlabel('$\ell_{1a,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_2p #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_2p' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{2,P}$', fontsize=30) ax.set_xlabel('$\ell_{2,p}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Ell_3s #--- --- --- --- --- #hyper-paramter name name_hyp = 'ell_3s' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 0.1 ymax_mean = 0.1 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\ell_{3,S}$', fontsize=30) ax.set_xlabel('$\ell_{3,s}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,150]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Tau_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'tau_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 150 ymax_mean = 150 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\tau_{0}$', fontsize=30) ax.set_xlabel(r'$\tau_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.5]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Phi_0 #--- --- --- --- --- #hyper-paramter name name_hyp = 'phi_0' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1000 ymax_mean = 1000 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\phi_{0}$', fontsize=30) ax.set_xlabel('$\phi_{0}$', fontsize=25) ax.set_ylabel(r'probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.6]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Omega_ca #--- --- --- --- --- #hyper-paramter name name_hyp = 'omega_cap' #figure title fname_fig = 'post_dist_' + name_hyp #create figure fig, ax = plt.subplots(figsize = (10,10)) for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): #estimate vertical line height for mean and mode ymax_mode = 1500 ymax_mean = 1500 #plot posterior dist pl_hyp = ax.vlines(df_r_h.loc['mean',name_hyp], ymin=0, ymax=ymax_mean, linestyle='-', label='Mean') ax.vlines(df_r_h.loc['prc_0.50',name_hyp], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_hyp.get_color(), label='Mode') #plot true value ymax_hyp = ymax_mean ax.vlines(np.sqrt(hyp['omega_ca1p']**2+hyp['omega_ca2p']**2), ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') #edit figure if not flag_report: ax.set_title(r'Comparison $\omega_{ca,P}$', fontsize=30) ax.set_xlabel('$\omega_{ca,p}$', fontsize=25) ax.set_ylabel('probability density function ', fontsize=25) ax.grid(which='both') ax.tick_params(axis='x', labelsize=22) ax.tick_params(axis='y', labelsize=22) #plot limits ax.set_xlim([0,0.05]) ax.set_ylim([0,ymax_hyp]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # # Delta c_0 # #--- --- --- --- --- # #hyper-paramter name # name_hyp = 'dc_0' # #figure title # fname_fig = 'post_dist_' + name_hyp # #create figure # fig, ax = plt.subplots(figsize = (10,10)) # for d_id, df_r_h, df_r_h_p in zip(ds_id, df_reg_hyp, df_reg_hyp_post): # #estimate vertical line height for mean and mode # ymax_mode = df_r_h_p.loc[:,name_hyp+'_pdf'].max() # ymax_mean = 1.5*np.ceil(ymax_mode/10)*10 # ymax_mean = 15 # #plot posterior dist # pl_pdf = ax.plot(df_r_h_p.loc[:,name_hyp], df_r_h_p.loc[:,name_hyp+'_pdf']) # ax.vlines(df_r_h.loc[name_hyp,'mean'], ymin=0, ymax=ymax_mean, linestyle='-', color=pl_pdf[0].get_color(), label='Mean') # ax.vlines(df_r_h.loc[name_hyp,'mode'], ymin=0, ymax=ymax_mode, linestyle='--', color=pl_pdf[0].get_color(), label='Mode') # #plot true value # ymax_hyp = ymax_mean # # ax.vlines(hyp[name_hyp], ymin=0, ymax=ymax_hyp, linestyle='-', linewidth=4, color='black', label='True value') # #edit figure # ax.set_title(r'Comparison $\delta c_{0}$', fontsize=30) # ax.set_xlabel('$\delta c_{0}$', fontsize=25) # ax.set_ylabel('probability density function ', fontsize=25) # ax.grid(which='both') # ax.tick_params(axis='x', labelsize=22) # ax.tick_params(axis='y', labelsize=22) # #plot limits # ax.set_xlim([-1,1]) # ax.set_ylim([0,ymax_hyp]) # #save figure # fig.tight_layout() # # fig.savefig( dir_fig + fname_fig + '.png' )
47,186
38.686291
153
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_pystan_model3_corr_cells_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model3_corr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Validation/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model3_corr_cells_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Validation/regression/ds3/' #output sub-directory out_dir_sub = 'PYSTAN_NGAWest2CANorth_corr_cells_chol_eff' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,635
29.049587
98
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/regression/ds3/main_pystan_model3_uncorr_cells_NGAWest2CANorth.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Jul 14 14:17:52 2021 @author: glavrent """ # Working directory and Packages # --------------------------- #load libraries import os import sys import numpy as np import pandas as pd import time #user functions sys.path.insert(0,'../../../Python_lib/regression/pystan/') from regression_pystan_model3_uncorr_cells_unbounded_hyp import RunStan # Define variables # --------------------------- #filename suffix # synds_suffix = '_small_corr_len' # synds_suffix = '_large_corr_len' #synthetic datasets directory ds_dir = '../../../../Data/Validation/synthetic_datasets/ds3' ds_dir = r'%s%s/'%(ds_dir, synds_suffix) # dataset info #ds_fname_main = 'CatalogNGAWest3CA_synthetic_data' ds_fname_main = 'CatalogNGAWest3CALite_synthetic_data' ds_id = np.arange(1,6) #cell specific anelastic attenuation ds_fname_cellinfo = 'CatalogNGAWest3CALite_cellinfo' ds_fname_celldist = 'CatalogNGAWest3CALite_distancematrix' #stan model sm_fname = '../../../Stan_lib/regression_stan_model3_uncorr_cells_unbounded_hyp_chol_efficient.stan' #output info #main output filename out_fname_main = 'NGAWest2CANorth_syndata' #main output directory out_dir_main = '../../../../Data/Validation/regression/ds3/' #output sub-directory out_dir_sub = 'PYSTAN_NGAWest2CANorth_uncorr_cells_chol_eff' #stan parameters runstan_flag = True # pystan_ver = 2 pystan_ver = 3 res_name = 'tot' n_iter = 1000 n_chains = 4 adapt_delta = 0.8 max_treedepth = 10 #ergodic coefficients c_2_erg=-2.0 c_3_erg=-0.6 c_a_erg=0.0 #parallel options # flag_parallel = True flag_parallel = False #output sub-dir with corr with suffix info out_dir_sub = f'%s%s'%(out_dir_sub, synds_suffix) #load cell dataframes cellinfo_fname = '%s%s.csv'%(ds_dir, ds_fname_cellinfo) celldist_fname = '%s%s.csv'%(ds_dir, ds_fname_celldist) df_cellinfo = pd.read_csv(cellinfo_fname) df_celldist = pd.read_csv(celldist_fname) # Run stan regression # --------------------------- #create datafame with computation time df_run_info = list() #iterate over all synthetic datasets for d_id in ds_id: print('Synthetic dataset %i fo %i'%(d_id, len(ds_id))) #run time start run_t_strt = time.time() #input flatfile ds_fname = '%s%s%s_Y%i.csv'%(ds_dir, ds_fname_main, synds_suffix, d_id) #load flatfile df_flatfile = pd.read_csv(ds_fname) #keep only North records of NGAWest2 df_flatfile = df_flatfile.loc[np.logical_and(df_flatfile.dsid==0, df_flatfile.sreg==1),:] #output file name and directory out_fname = '%s%s_Y%i'%(out_fname_main, synds_suffix, d_id) out_dir = '%s/%s/Y%i/'%(out_dir_main, out_dir_sub, d_id) #run stan model RunStan(df_flatfile, df_cellinfo, df_celldist, sm_fname, out_fname, out_dir, res_name, c_2_erg=c_2_erg, c_3_erg=c_3_erg, c_a_erg=c_a_erg, runstan_flag=runstan_flag, n_iter=n_iter, n_chains=n_chains, adapt_delta=adapt_delta, max_treedepth=max_treedepth, pystan_ver=pystan_ver, pystan_parallel=flag_parallel) #run time end run_t_end = time.time() #compute run time run_tm = (run_t_end - run_t_strt)/60 #log run time df_run_info.append(pd.DataFrame({'computer_name':os.uname()[1],'out_name':out_dir_sub, 'ds_id':d_id,'run_time':run_tm}, index=[d_id])) #write out run info out_fname = '%s%s/run_info.csv'%(out_dir_main, out_dir_sub) pd.concat(df_run_info).reset_index(drop=True).to_csv(out_fname, index=False)
3,639
29.082645
100
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/preprocessing/CreateCatalogNGAWest2CA.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Jun 27 22:58:16 2021 @author: glavrent """ # %% Required Packages # ====================================== #load libraries import os import sys import pathlib import glob import re #regular expression package #arithmetic libraries import numpy as np import pandas as pd #geographic coordinates import pyproj #plotting libraries from matplotlib import pyplot as plt import matplotlib.ticker as mticker #user-derfined functions sys.path.insert(0,'../../Python_lib/catalog') sys.path.insert(0,'../../Python_lib/plotting') import pylib_catalog as pylib_catalog import pylib_contour_plots as pylib_cplt # %% Define Input Data # ====================================== #thresholds thres_dist = 0.01 #collocated stations # projection system utm_zone = '11S' # region id region_id = 1 #input file names fname_flatfile_NGA2ASK14 = '../../../Raw_files/nga_w2_resid/resid_T0.200.out2.txt' fname_flatfile_NGA2coor = '../../../Raw_files/nga_w2/Updated_NGA_West2_Flatfile_coordinates.csv' #flatfile file fname_flatfile = 'CatalogNGAWest2CA_ASK14' #output directory dir_out = '../../../Data/Verification/preprocessing/flatfiles/NGAWest2_CA/' dir_fig = dir_out + 'figures/' # %% Load Data # ====================================== #NGAWest2 df_flatfile_NGA2ASK14 = pd.read_csv(fname_flatfile_NGA2ASK14, delim_whitespace=True) df_flatfile_NGA2coor = pd.read_csv(fname_flatfile_NGA2coor) df_flatfile_NGA2 = pd.merge(df_flatfile_NGA2ASK14, df_flatfile_NGA2coor, left_on='recID', right_on='Record Sequence Number') # %% Cleaning files # ====================================== # NGA2 #keep only CA for NGA2 df_flatfile_NGA2 = df_flatfile_NGA2[ df_flatfile_NGA2.region == region_id ] #reset indices df_flatfile_NGA2.reset_index(inplace=True) # %% Process Data # ====================================== #coordinates and projection system # projection system utmProj = pyproj.Proj("+proj=utm +zone="+utm_zone+", +ellps=WGS84 +datum=WGS84 +units=m +no_defs") #earthquake and station ids eq_id_NGA2 = df_flatfile_NGA2['eqid'].values.astype(int) sta_id_NGA2 = df_flatfile_NGA2['Station Sequence Number'].values.astype(int) #unique earthquake and station ids eq_id_NGA2_unq, eq_idx_NGA2 = np.unique(eq_id_NGA2, return_index=True) sta_id_NGA2_unq, sta_idx_NGA2 = np.unique(sta_id_NGA2, return_index=True) #number of earthquake and stations neq_NGA2 = len(eq_id_NGA2_unq) nsta_NGA2 = len(sta_id_NGA2_unq) #earthquake and station coordinates eq_latlon_NGA2_all = df_flatfile_NGA2[['Hypocenter Latitude (deg)','Hypocenter Longitude (deg)']].values sta_latlon_NGA2_all = df_flatfile_NGA2[['Station Latitude','Station Longitude']].values #utm coordinates eq_X_NGA2_all = np.array([utmProj(e_lon, e_lat) for e_lat, e_lon in zip(eq_latlon_NGA2_all[:,0], eq_latlon_NGA2_all[:,1])]) / 1000 eq_z_NGA2_all = -1*df_flatfile_NGA2['Hypocenter Depth (km)'].values sta_X_NGA2_all = np.array([utmProj(s_lon, s_lat) for s_lat, s_lon in zip(sta_latlon_NGA2_all[:,0], sta_latlon_NGA2_all[:,1])]) / 1000 mpt_X_NGA2_all = (eq_X_NGA2_all + sta_X_NGA2_all) / 2 #mid point coordinates mpt_latlon_NGA2_all = np.flip( np.array([utmProj(pt_x, pt_y, inverse=True) for pt_x, pt_y in zip(mpt_X_NGA2_all[:,0], mpt_X_NGA2_all[:,1]) ]), axis=1) #ground motion parameteres mag_NGA2 = df_flatfile_NGA2['mag'].values rup_NGA2 = np.sqrt(np.linalg.norm(eq_X_NGA2_all-sta_X_NGA2_all, axis=1)**2+eq_z_NGA2_all**2) vs30_NGA2 = df_flatfile_NGA2['VS30'].values # %% Process Data to save # ====================================== i_data2keep = np.full(len(df_flatfile_NGA2), True) #records' info rsn_array = df_flatfile_NGA2['recID'].values[i_data2keep].astype(int) eqid_array = eq_id_NGA2[i_data2keep] ssn_array = sta_id_NGA2[i_data2keep] year_array = df_flatfile_NGA2['YEAR'].values[i_data2keep] #records' parameters mag_array = mag_NGA2[i_data2keep] rrup_array = rup_NGA2[i_data2keep] vs30_array = vs30_NGA2[i_data2keep] #earthquake, station, mid-point latlon coordinates eq_latlon = eq_latlon_NGA2_all[i_data2keep,:] sta_latlon = sta_latlon_NGA2_all[i_data2keep,:] mpt_latlon = mpt_latlon_NGA2_all[i_data2keep,:] #earthquake, station, mid-point UTM coordinates eq_utm = eq_X_NGA2_all[i_data2keep,:] sta_utm = sta_X_NGA2_all[i_data2keep,:] mpt_utm = mpt_X_NGA2_all[i_data2keep,:] #earthquake source depth eq_z = eq_z_NGA2_all[i_data2keep] #indices for unique earthquakes and stations _, eq_idx, eq_inv = np.unique(eqid_array, return_index=True, return_inverse=True) _, sta_idx, sta_inv = np.unique(ssn_array, return_index=True, return_inverse=True) n_eq_orig = len(eq_idx) n_sta_orig = len(sta_idx) # NGAWest2 dataframe # ---- ---- ---- ---- ---- data_full = {'rsn':rsn_array, 'eqid':eqid_array, 'ssn':ssn_array, 'mag':mag_array, 'Rrup':rrup_array, 'Vs30': vs30_array, 'year': year_array, 'eqLat':eq_latlon[:,0], 'eqLon':eq_latlon[:,1], 'staLat':sta_latlon[:,0], 'staLon':sta_latlon[:,1], 'mptLat':mpt_latlon[:,0], 'mptLon':mpt_latlon[:,1], 'UTMzone':utm_zone, 'eqX':eq_utm[:,0], 'eqY':eq_utm[:,1], 'eqZ':eq_z, 'staX':sta_utm[:,0], 'staY':sta_utm[:,1], 'mptX':mpt_utm[:,0], 'mptY':mpt_utm[:,1]} df_flatfile_full = pd.DataFrame(data_full) # colocate stations # ---- ---- ---- ---- ---- #update ssn for colocated stations df_flatfile_full = pylib_catalog.ColocatePt(df_flatfile_full, 'ssn', ['staX','staY'], thres_dist=thres_dist) #keep single record from each event after collocation # ---- ---- ---- ---- ---- i_unq_eq_sta = np.unique(df_flatfile_full[['eqid','ssn']].values, return_index=True, axis=0)[1] df_flatfile_full = df_flatfile_full.iloc[i_unq_eq_sta, :].sort_index() _, eq_idx, eq_inv = np.unique(df_flatfile_full.loc[:,'eqid'], axis=0, return_index=True, return_inverse=True) _, sta_idx, sta_inv = np.unique(df_flatfile_full.loc[:,'ssn'], axis=0, return_index=True, return_inverse=True) n_eq = len(eq_idx) n_sta = len(sta_idx) # average gm parameters # ---- ---- ---- ---- ---- df_flatfile_full = pylib_catalog.IndexAvgColumns(df_flatfile_full, 'eqid', ['mag','eqLat','eqLon','eqX','eqY','eqZ']) df_flatfile_full = pylib_catalog.IndexAvgColumns(df_flatfile_full, 'ssn', ['Vs30','staLat','staLon','staX','staY']) # create event and station dataframes # ---- ---- ---- ---- ---- #event dataframe df_flatfile_event = df_flatfile_full.iloc[eq_idx,:][['eqid','mag','year','eqLat','eqLon','UTMzone','eqX','eqY','eqZ']].reset_index(drop=True) #station dataframe df_flatfile_station = df_flatfile_full.iloc[sta_idx,:][['ssn','Vs30','staLat','staLon','UTMzone','staX','staY']].reset_index(drop=True) # %% Save data # ====================================== # create output directories if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) if not os.path.isdir(dir_fig): pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save processed dataframes fname_flatfile_full= '%s%s'%(dir_out, fname_flatfile) df_flatfile_full.to_csv(fname_flatfile_full + '.csv', index=False) df_flatfile_event.to_csv(fname_flatfile_full + '_event.csv', index=False) df_flatfile_station.to_csv(fname_flatfile_full + '_station.csv', index=False) # create figures # ---- ---- ---- ---- ---- # Mag-Dist distribution fname_fig = 'M-R_dist' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(df_flatfile_full.Rrup, df_flatfile_full.mag, label='NGAWest2 CA') #edit figure properties ax.set_xlabel(r'Distance ($km$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') ax.set_xscale('log') # ax.set_xlim([0.1, 2000]) ax.set_ylim([2, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) # ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' ) # Mag-Year distribution fname_fig = 'M-date_dist' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(df_flatfile_event['year'].values, df_flatfile_event['mag'].values, label='NGAWest2 CA') #edit figure properties ax.set_xlabel(r'time ($year$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') # ax.set_xscale('log') ax.set_xlim([1965, 2025]) ax.set_ylim([2, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) # ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' ) #eq and sta location fname_fig = 'eq_sta_locations' fig, ax, data_crs, gl = pylib_cplt.PlotMap(flag_grid=True) #plot earthquake and station locations ax.plot(df_flatfile_event['eqLon'].values, df_flatfile_event['eqLat'].values, '*', transform = data_crs, markersize = 10, zorder=13, label='Events') ax.plot(df_flatfile_station['staLon'].values, df_flatfile_station['staLat'].values, 'o', transform = data_crs, markersize = 6, zorder=12, label='Stations') #edit figure properties gl.xlabel_style = {'size': 25} gl.ylabel_style = {'size': 25} # gl.xlocator = mticker.FixedLocator([-124, -122, -120, -118, -116, -114]) # gl.ylocator = mticker.FixedLocator([32, 34, 36, 38, 40]) ax.legend(fontsize=25, loc='lower left') # ax.set_xlim(plt_latlon_win[:,1]) # ax.set_ylim(plt_latlon_win[:,0]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Print data info # --------------------------- print(f'NGAWest2:') print(f'\tnumber of rec: %.i'%len(df_flatfile_full)) print(f'\tnumber of rec (R<200km): %.i'%np.sum(df_flatfile_full.Rrup<=200)) print(f'\tnumber of rec (R<300km): %.i'%np.sum(df_flatfile_full.Rrup<=300)) print(f'\tnumber of eq: %.i'%len(df_flatfile_event)) print(f'\tnumber of sta: %.i'%len(df_flatfile_station)) print(f'\tcoverage: %.i to %i'%(df_flatfile_full.year.min(), df_flatfile_full.year.max())) #write out summary # ---- ---- ---- ---- ---- f = open(dir_out + 'summary_data' + '.txt', 'w') f.write(f'NGAWest2:\n') f.write(f'\tnumber of rec: %.i\n'%len(df_flatfile_full)) f.write(f'\tnumber of rec (R<200km): %.i\n'%np.sum(df_flatfile_full.Rrup<=200)) f.write(f'\tnumber of rec (R<300km): %.i\n'%np.sum(df_flatfile_full.Rrup<=300)) f.write(f'\tnumber of eq: %.i\n'%len(df_flatfile_event)) f.write(f'\tnumber of sta: %.i\n'%len(df_flatfile_station)) f.write(f'\tcoverage: %.i to %i\n'%(df_flatfile_full.year.min(), df_flatfile_full.year.max())) f.close()
11,132
39.631387
164
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/preprocessing/CreateCatalogNewEvents2017.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Jun 27 16:12:57 2021 @author: glavrent """ # Required Packages # ====================================== #load libraries import os import sys import pathlib import glob import re #regular expression package #arithmetic libraries import numpy as np import pandas as pd #geographic coordinates import pyproj #plotting libraries from matplotlib import pyplot as plt import matplotlib.ticker as mticker #user-derfined functions sys.path.insert(0,'../../Python_lib/ground_motions') sys.path.insert(0,'../../Python_lib/plotting') import pylib_Willis15CA_Vs30 as pylib_W15_Vs30 import pylib_contour_plots as pylib_cplt # %% Define Input Data # ====================================== #thresholds dt_thres = 0.025 rrup_thres = 300 # projection system utm_zone = '11S' #input flatfiles fname_flatfile_newrec_eq = '../../../Raw_files/nga_w3/California2011-2017/eventcatalog.csv' fname_flatfile_newrec_sta = '../../../Raw_files/nga_w3/California2011-2017/Recorddata.csv' #flatfile file fname_flatfile = 'CatalogNewRecords_2011-2017_CA_NV' #output directory dir_out = '../../../Data/NGAWest_expansion/CA_NV_2011-2017/' dir_out = '../../../Data/Verification/preprocessing/flatfiles/CA_NV_2011-2017/' dir_fig = dir_out + 'figures/' # %% Load Data # ====================================== #merge event and station info df_flatfile_newrec_eq = pd.read_csv(fname_flatfile_newrec_eq) df_flatfile_newrec_sta = pd.read_csv(fname_flatfile_newrec_sta) df_flatfile_newrec = pd.merge(df_flatfile_newrec_eq, df_flatfile_newrec_sta, left_on='EventIDs.i.', right_on='EventID') # %% Cleaning files # ====================================== #set -999 to nan df_flatfile_newrec.replace(-999, np.nan, inplace=True) #remove data based on timestep df_flatfile_newrec = df_flatfile_newrec[ df_flatfile_newrec.timeStep <= dt_thres ] #remove data with unknown mag df_flatfile_newrec = df_flatfile_newrec[ ~np.isnan(df_flatfile_newrec['mag.event']) ] #remove data with unknown coordinates df_flatfile_newrec = df_flatfile_newrec[ ~np.isnan(df_flatfile_newrec[['latitude.event', 'longitude.event']]).any(axis=1) ] df_flatfile_newrec = df_flatfile_newrec[ ~np.isnan(df_flatfile_newrec[['stnlat', 'stnlon']]).any(axis=1) ] #keep single record from df_flatfile_newrec.loc[:,'EventID'] = df_flatfile_newrec['EventID'].values.astype(int) df_flatfile_newrec.loc[:,'stationID'] = np.unique(df_flatfile_newrec['station'], return_inverse=True)[1] i_unq_eq_sta = np.unique( np.unique(df_flatfile_newrec[['stationID','EventID']].values, return_index=True, axis=0)[1] ) df_flatfile_newrec = df_flatfile_newrec.iloc[i_unq_eq_sta, :] #reset indices df_flatfile_newrec.reset_index(inplace=True) # %% Process Data # ====================================== #coordinates and projection system # projection system utmProj = pyproj.Proj("+proj=utm +zone="+utm_zone+", +ellps=WGS84 +datum=WGS84 +units=m +no_defs") #earthquake and station ids eq_id_newrec = df_flatfile_newrec['EventIDs.i.'].values.astype(int) sta_id_newrec = df_flatfile_newrec['station'].values sta_net_newrec = df_flatfile_newrec['network'].values sta_netid_newrec = [f'{s_net}-{s_id}' for s_net, s_id in zip(sta_net_newrec, sta_id_newrec)] #unique earthquake and station ids eq_id_newrec_unq, eq_idx_newrec = np.unique(eq_id_newrec, return_index=True) # sta_id_newrec_unq, sta_inv_newrec = np.unique(sta_id_newrec, return_inverse=True) sta_id_newrec_unq, sta_inv_newrec = np.unique(sta_netid_newrec, return_inverse=True) #number of earthquake and stations neq_newrec = len(eq_id_newrec_unq) nsta_newrec = len(sta_id_newrec_unq) #earthquake and station coordinates eq_latlon_newrec_all = df_flatfile_newrec[['latitude.event', 'longitude.event']].values sta_latlon_newrec_all = df_flatfile_newrec[['stnlat', 'stnlon']].values #utm coordinates eq_X_newrec_all = np.array([utmProj(e_lon, e_lat) for e_lat, e_lon in zip(eq_latlon_newrec_all[:,0], eq_latlon_newrec_all[:,1])]) / 1000 eq_z_newrec_all = np.minimum(-1*df_flatfile_newrec['depth.event.1000'].values, 0) sta_X_newrec_all = np.array([utmProj(s_lon, s_lat) for s_lat, s_lon in zip(sta_latlon_newrec_all[:,0], sta_latlon_newrec_all[:,1])]) / 1000 mpt_X_newrec_all = (eq_X_newrec_all + sta_X_newrec_all) / 2 #mid point coordinates mpt_latlon_newrec_all = np.flip( np.array([utmProj(pt_x, pt_y, inverse=True) for pt_x, pt_y in zip(mpt_X_newrec_all[:,0], mpt_X_newrec_all[:,1]) ]), axis=1 ) #earthquake parameteres mag_newrec = df_flatfile_newrec['mag.event'].values rup_newrec = np.sqrt(np.linalg.norm(eq_X_newrec_all-sta_X_newrec_all, axis=1)**2+eq_z_newrec_all**2) #year of recording df_flatfile_newrec['year'] = pd.DatetimeIndex(df_flatfile_newrec['rec.stime']).year #estimate station vs30 Wills15Vs30 = pylib_W15_Vs30.Willis15Vs30CA() vs30_newrec = Wills15Vs30.lookup(np.fliplr(sta_latlon_newrec_all))[0] vs30_newrec[vs30_newrec<=50] = np.nan # %% Process Data to save # ====================================== #distance threshold for data to keep i_data2keep = rup_newrec <= rrup_thres #records' info rsn_array = df_flatfile_newrec.loc[i_data2keep,'index'].values eqid_array = eq_id_newrec[i_data2keep] ssn_array = sta_inv_newrec[i_data2keep] sid_array = sta_id_newrec[i_data2keep] snet_array = sta_net_newrec[i_data2keep] year_array = df_flatfile_newrec['year'].values[i_data2keep] #records' parameters mag_array = mag_newrec[i_data2keep] rrup_array = rup_newrec[i_data2keep] vs30_array = vs30_newrec[i_data2keep] #earthquake, station, mid-point latlon coordinates eq_latlon = eq_latlon_newrec_all[i_data2keep,:] sta_latlon = sta_latlon_newrec_all[i_data2keep,:] mpt_latlon = mpt_latlon_newrec_all[i_data2keep,:] #earthquake, station, mid-point UTM coordinates eq_utm = eq_X_newrec_all[i_data2keep,:] sta_utm = sta_X_newrec_all[i_data2keep,:] mpt_utm = mpt_X_newrec_all[i_data2keep,:] #earthquake source depth eq_z = eq_z_newrec_all[i_data2keep] #indices for unique earthquakes and stations eq_idx = np.unique(eqid_array, return_index=True)[1] sta_idx = np.unique(ssn_array, return_index=True)[1] #data to save data_full = {'rsn':rsn_array, 'eqid':eqid_array, 'ssn':ssn_array, 'mag':mag_array, 'Rrup':rrup_array, 'Vs30': vs30_array, 'year': year_array, 'eqLat':eq_latlon[:,0], 'eqLon':eq_latlon[:,1], 'staLat':sta_latlon[:,0], 'staLon':sta_latlon[:,1], 'mptLat':mpt_latlon[:,0], 'mptLon':mpt_latlon[:,1], 'UTMzone':utm_zone, 'eqX':eq_utm[:,0], 'eqY':eq_utm[:,1], 'eqZ':eq_z, 'staX':sta_utm[:,0], 'staY':sta_utm[:,1], 'mptX':mpt_utm[:,0], 'mptY':mpt_utm[:,1]} #processed dataframes df_flatfile_full = pd.DataFrame(data_full) #event dataframe df_flatfile_event = df_flatfile_full.loc[eq_idx, ['eqid','mag','year','eqLat','eqLon','UTMzone','eqX','eqY','eqZ']].reset_index(drop=True) #station dataframe df_flatfile_station = df_flatfile_full.loc[sta_idx, ['ssn','Vs30','staLat','staLon','UTMzone','staX','staY']].reset_index(drop=True) # %% Save data # ====================================== # create output directories if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) if not os.path.isdir(dir_fig): pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save processed dataframes fname_flatfile_full= '%s%s'%(dir_out, fname_flatfile) df_flatfile_full.to_csv(fname_flatfile_full + '.csv', index=True) df_flatfile_event.to_csv(fname_flatfile_full + '_event.csv', index=False) df_flatfile_station.to_csv(fname_flatfile_full + '_station.csv', index=False) # create figures # ---- ---- ---- ---- ---- # Mag-Dist distribution fname_fig = 'M-R_dist' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(df_flatfile_full.Rrup, df_flatfile_full.mag, label='New Records') #edit figure properties ax.set_xlabel(r'Distance ($km$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') ax.set_xscale('log') # ax.set_xlim([0.1, 2000]) ax.set_ylim([2, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' ) # Mag-Year distribution fname_fig = 'M-date_dist' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(df_flatfile_event['year'].values, df_flatfile_event['mag'].values, label='New Records') #edit figure properties ax.set_xlabel(r'time ($year$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') # ax.set_xscale('log') ax.set_xlim([1965, 2025]) ax.set_ylim([2, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' ) #eq and sta location fname_fig = 'eq_sta_locations' fig, ax, data_crs, gl = pylib_cplt.PlotMap(flag_grid=True) #plot earthquake and station locations ax.plot(df_flatfile_event['eqLon'].values, df_flatfile_event['eqLat'].values, '*', transform = data_crs, markersize = 10, zorder=13) ax.plot(df_flatfile_station['staLon'].values, df_flatfile_station['staLat'].values, 'o', transform = data_crs, markersize = 6, zorder=13, label='STA') ax.plot(df_flatfile_event['eqLon'].values, df_flatfile_event['eqLat'].values, '*', color='black', transform = data_crs, markersize = 14, zorder=13, label='EQ') #edit figure properties gl.xlabel_style = {'size': 25} gl.ylabel_style = {'size': 25} # gl.xlocator = mticker.FixedLocator([-124, -122, -120, -118, -116, -114]) # gl.ylocator = mticker.FixedLocator([32, 34, 36, 38, 40]) ax.legend(fontsize=25, loc='lower left') # ax.set_xlim(plt_latlon_win[:,1]) # ax.set_ylim(plt_latlon_win[:,0]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Print data info # --------------------------- print(r'New Records:') print(f'\tnumber of rec: %.i'%len(df_flatfile_full)) print(f'\tnumber of rec (R<200km): %.i'%np.sum(df_flatfile_full.Rrup<=200)) print(f'\tnumber of rec (R<%.1f): %.i'%(rrup_thres, np.sum(df_flatfile_full.Rrup<=rrup_thres))) print(f'\tnumber of eq: %.i'%len(df_flatfile_event)) print(f'\tnumber of sta: %.i'%len(df_flatfile_station)) print(f'\tnumber of sta (R<300km): %.i'%len(np.unique(ssn_array[df_flatfile_full.Rrup<=300]))) print(f'\tcoverage: %.i to %i'%(df_flatfile_full.year.min(), df_flatfile_full.year.max()))
11,196
41.25283
165
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/preprocessing/CreateCatalogNewEvents2021Raw.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Oct 5 09:37:23 2021 @author: glavrent """ # %% Required Packages # ====================================== #load libraries import os import sys import pathlib import re #arithmetic libraries import numpy as np import pandas as pd #geographical libraries import geopy from geopy.distance import distance #plotting libraries from matplotlib import pyplot as plt import matplotlib.ticker as mticker #user-derfined functions sys.path.insert(0,'../../Python_lib/catalog') sys.path.insert(0,'../../Python_lib/plotting') import pylib_catalog as pylib_catalog import pylib_contour_plots as pylib_cplt # %% Define variables # ====================================== #input file names fname_nrec_eq = '../../../Raw_files/nga_w3/IRIS/fdsnws-events_CA_2011-2021.csv' fname_nerc_sta = '../../../Raw_files/nga_w3/IRIS/fdsn-station_USA_[HB][HN]?.csv' fname_mag_rup_lim = '../../../Data/Verification/preprocessing/flatfiles/usable_mag_rrup/usable_Mag_Rrup_coeffs.csv' #output directoy dir_out = '../../../Data/Verification/preprocessing/flatfiles/CA_NV_2011-2021_raw/' dir_fig = dir_out + 'figures/' fname_new_cat = 'Catalog_California_2011-2021.ver02' #station srate_min = 10 #minimum sample rate for accepting instrument net_seis = np.array(['AZ','BK','CI','NC','NN','NP','PB','PG','SB','WR','US','CJ','UO']) # %% Load Data # ====================================== df_nrec_eq = pd.read_csv(fname_nrec_eq, delimiter='|', skiprows=4) df_nrec_sta = pd.read_csv(fname_nerc_sta, delimiter='|', skiprows=4) #M/R limitis df_lim_mag_rrup = pd.read_csv(fname_mag_rup_lim, index_col=0) # %% Process Data # ====================================== #rename earthquake and station coordinates df_nrec_eq.rename(columns={'Latitude':'eqLat', 'Longitude':'eqLon', 'Depth':'eqDepth'}, inplace=True) df_nrec_sta.rename(columns={'Latitude':'staLat', 'Longitude':'staLon','Depth':'staDepth'}, inplace=True) #remove rows with nna columns df_nrec_eq = df_nrec_eq.loc[~df_nrec_eq.isnull().any(axis=1).values,:] df_nrec_sta = df_nrec_sta.loc[~df_nrec_sta.isnull().any(axis=1).values,:] df_nrec_eq = df_nrec_eq.loc[~df_nrec_eq.isna().any(axis=1).values,:] df_nrec_sta = df_nrec_sta.loc[~df_nrec_sta.isna().any(axis=1).values,:] #remove invalid columns i_nan_rec = np.array([bool(re.match('^#.*$',n)) for n in df_nrec_sta['Network']]) df_nrec_sta = df_nrec_sta.loc[~i_nan_rec,:] #keep networks of interest i_net = df_nrec_sta.Network.isin(net_seis) df_nrec_sta = df_nrec_sta.loc[i_net,:] #with only records with sufficient sampling rate df_nrec_sta.SampleRate = df_nrec_sta.SampleRate.astype(float) i_srate = df_nrec_sta.SampleRate > srate_min df_nrec_sta = df_nrec_sta.loc[i_srate,:] #create network and station IDs _, df_nrec_sta.loc[:,'NetworkID'] = np.unique(df_nrec_sta['Network'].values.astype(str), return_inverse=True) _, df_nrec_sta.loc[:,'StationID'] = np.unique(df_nrec_sta['Station'].values.astype(str), return_inverse=True) #reduce to unique stations _, i_sta_unq = np.unique(df_nrec_sta[['NetworkID','StationID']], return_index=True, axis=0) df_nrec_sta = df_nrec_sta.iloc[i_sta_unq,:] #station coordinates sta_latlon = df_nrec_sta[['staLat','staLon']].values #initialize rec catalog cat_new_rec = [] #number of events and stations n_eq = len(df_nrec_eq) n_sta = len(df_nrec_sta) #iterate evetns for (k, eq) in df_nrec_eq.iterrows(): print('Processing event %i of %i'%(k, n_eq)) #earthquake info eq_latlon = eq[['eqLat','eqLon']].values eq_depth = eq['eqDepth'] eq_mag = eq['Magnitude'] #epicenter and hypocenter dist dist_epi = np.array([distance(eq_latlon, sta_ll).km for sta_ll in sta_latlon]) dist_hyp = np.sqrt(dist_epi**2 + eq_depth**2) #stations that satisfy the M/R limit i_sta_event = pylib_catalog.UsableSta(np.full(n_sta, eq_mag), dist_hyp, df_lim_mag_rrup) #create catalog for k^th event df_new_r = df_nrec_sta.loc[i_sta_event,:].assign(**eq) #add rupture info df_new_r.loc[:,'HypDist'] = dist_hyp[i_sta_event] #combine sta with event info cat_new_rec.append(df_new_r) #combine catalogs of all events into one df_cat_new_rec = pd.concat(cat_new_rec).reset_index() #re-roder columns df_cat_new_rec = df_cat_new_rec[np.concatenate([df_nrec_eq.columns, df_nrec_sta.columns,['HypDist']])] #create event and station dataframes #indices for unique earthquakes and stations eq_idx = np.unique(df_cat_new_rec.EventID, return_index=True)[1] sta_idx = np.unique(df_cat_new_rec[['NetworkID','StationID']], return_index=True, axis=0)[1] #event dataframe df_cat_new_rec_eq = df_cat_new_rec.loc[eq_idx, df_nrec_eq.columns].reset_index(drop=True) #station dataframe df_cat_new_rec_sta = df_cat_new_rec.loc[sta_idx, df_nrec_sta.columns].reset_index(drop=True) # %% Output # ====================================== # create output directories if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) if not os.path.isdir(dir_fig): pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #save processed dataframes fname_cat = '%s%s'%(dir_out, fname_new_cat ) df_cat_new_rec.to_csv(fname_cat + '.csv', index=False) df_cat_new_rec_eq.to_csv(fname_cat + '_event.csv', index=False) df_cat_new_rec_sta.to_csv(fname_cat + '_station.csv', index=False) # create figures # ---- ---- ---- ---- ---- # Mag-Dist distribution fname_fig = 'M-R_dist' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(df_cat_new_rec.HypDist, df_cat_new_rec.Magnitude) #edit figure properties ax.set_xlabel(r'Distance ($km$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') # ax.set_xscale('log') # ax.set_xlim([0.1, 2000]) ax.set_ylim([1, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) # ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' ) #log scale figure ax.set_xscale('log') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '_log' + '.png' ) # Mag-Year distribution fname_fig = 'M-date_dist' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(pd.DatetimeIndex(df_cat_new_rec['Time']).year, df_cat_new_rec['Magnitude'].values) #edit figure properties ax.set_xlabel(r'time ($year$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') # ax.set_xscale('log') ax.set_xlim([1965, 2025]) ax.set_ylim([1, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) # ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' ) #eq and sta location fname_fig = 'eq_sta_locations' fig, ax, data_crs, gl = pylib_cplt.PlotMap(flag_grid=True) #plot earthquake and station locations ax.plot(df_cat_new_rec_eq['eqLon'].values, df_cat_new_rec_eq['eqLat'].values, '*', transform = data_crs, markersize = 10, zorder=13, label='Events') ax.plot(df_cat_new_rec_sta['staLon'].values, df_cat_new_rec_sta['staLat'].values, 'o', transform = data_crs, markersize = 6, zorder=13, label='Stations') #edit figure properties gl.xlabel_style = {'size': 25} gl.ylabel_style = {'size': 25} # gl.xlocator = mticker.FixedLocator([-124, -122, -120, -118, -116, -114]) # gl.ylocator = mticker.FixedLocator([32, 34, 36, 38, 40]) ax.legend(fontsize=25, loc='lower left') # ax.set_xlim(plt_latlon_win[:,1]) # ax.set_ylim(plt_latlon_win[:,0]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # Print data info # --------------------------- print(f'New Records:') print(f'\tnumber of rec: %.i'%len(df_cat_new_rec)) print(f'\tnumber of rec (R<200km): %.i'%np.sum(df_cat_new_rec.HypDist<=200)) print(f'\tnumber of rec (R<300km): %.i'%np.sum(df_cat_new_rec.HypDist<=300)) print(f'\tnumber of eq: %.i'%len(df_cat_new_rec_eq)) print(f'\tnumber of sta: %.i'%len(df_cat_new_rec_sta))
8,640
36.569565
154
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/preprocessing/CreateCatalogNewEvents2021.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Jun 27 16:12:57 2021 @author: glavrent """ # Required Packages # ====================================== #load libraries import os import sys import pathlib import glob import re #regular expression package #arithmetic libraries import numpy as np import pandas as pd #geographic coordinates import pyproj #plotting libraries from matplotlib import pyplot as plt import matplotlib.ticker as mticker #user-derfined functions sys.path.insert(0,'../../Python_lib/catalog') sys.path.insert(0,'../../Python_lib/plotting') import pylib_catalog as pylib_catalog import pylib_contour_plots as pylib_cplt # %% Define Input Data # ====================================== #thresholds #min sampling rate thres_dt = 0.025 #collocated stations thres_dist = 0.01 #maximum depth eq_mag_depth = 25 #distance range # rrup_thres = 300 rrup_thres = 400 #year range year_min = 2011 year_max = 2021 # year_max = 2013 # projection system utm_zone = '11S' #input flatfiles fname_flatfile_newrec = '../../../Data/Verification/preprocessing/flatfiles/CA_NV_2011-2021_raw/Catalog_California_2011-2021.ver02.csv' fname_sta_vs30 = '../../../Raw_files/nga_w3/IRIS/Catalog_California_2011-2021.ver02_station_Vs30PW.csv' #flatfile file fname_flatfile = 'CatalogNewRecords_%.i-%.i_CA_NV'%(year_min, year_max ) #output directory dir_out = '../../../Data/Verification/preprocessing/flatfiles/CA_NV_%.i-%.i/'%(year_min, year_max) dir_fig = dir_out + 'figures/' #latlon window win_latlon = np.array([[30, 43],[-125, -110]]) #win_latlon = np.array([[32, 43],[-125, -114]]) # %% Load Data # ====================================== #read event and station info df_flatfile_newrec = pd.read_csv(fname_flatfile_newrec) #read vs30 info if fname_sta_vs30: df_sta_vs30 = pd.read_csv(fname_sta_vs30) else: df_sta_vs30 = None # %% Process Data # ====================================== # projection system # ---- ---- ---- ---- ---- utmProj = pyproj.Proj("+proj=utm +zone="+utm_zone+", +ellps=WGS84 +datum=WGS84 +units=km +no_defs") # rename columns # ---- ---- ---- ---- ---- df_flatfile_newrec.rename(columns={'EventID':'eventid', 'Time':'time', 'Latitude':'eqLat', 'Longitude':'eqLon', 'Author':'author', 'Catalog':'cat', 'Contributor':'contributor', 'ContributorID':'contributor_id', 'Magnitude':'mag', 'MagType':'mag_type', 'MagAuthor':'mag_author', 'EventLocationName':'eq_loc', 'ESN':'esn', 'Double.event':'double_event', 'event.ESN':'esn_y', 'event.EventID':'eventid', 'Network':'network', 'Station':'station', 'Latitude':'staLat', 'Longitude':'staLon', 'Elevation':'staElev'}, inplace=True) #year of event df_flatfile_newrec['year'] = pd.DatetimeIndex(df_flatfile_newrec['time']).year #station vs30 if fname_sta_vs30: df_sta_vs30.rename(columns={'Network':'network', 'Station':'station', 'Vs30_lnSD': 'Vs30sig'}, inplace=True) _, df_sta_vs30.loc[:,'sta_id'] = np.unique(df_sta_vs30.station, return_inverse=True) _, df_sta_vs30.loc[:,'net_id'] = np.unique(df_sta_vs30.network, return_inverse=True) assert(len(df_sta_vs30) == len(np.unique(df_sta_vs30[['net_id','sta_id']], axis=1))),'Error. Non-unique network stations' # cleaning files # ---- ---- ---- ---- ---- #set -999 to nan df_flatfile_newrec.replace(-999, np.nan, inplace=True) #remove data with unknown mag df_flatfile_newrec = df_flatfile_newrec[ ~np.isnan(df_flatfile_newrec['mag']) ] #remove data with unknown coordinates df_flatfile_newrec = df_flatfile_newrec[ ~np.isnan(df_flatfile_newrec[['eqLat', 'eqLon']]).any(axis=1) ] df_flatfile_newrec = df_flatfile_newrec[ ~np.isnan(df_flatfile_newrec[['staLat', 'staLon']]).any(axis=1) ] # keep only data in spatio-temporal window # ---- ---- ---- ---- ---- #earthquakes i_space_win_eq = np.all(np.array([df_flatfile_newrec.eqLat >= win_latlon[0,0], df_flatfile_newrec.eqLat < win_latlon[0,1], df_flatfile_newrec.eqLon >= win_latlon[1,0], df_flatfile_newrec.eqLon < win_latlon[1,1]]),axis=0) #stations i_space_win_sta = np.all(np.array([df_flatfile_newrec.staLat >= win_latlon[0,0], df_flatfile_newrec.staLat < win_latlon[0,1], df_flatfile_newrec.staLon >= win_latlon[1,0], df_flatfile_newrec.staLon < win_latlon[1,1]]),axis=0) #depth limit i_eq_depth = df_flatfile_newrec.eqDepth <= eq_mag_depth #time i_time_win = np.logical_and(df_flatfile_newrec.year >= year_min, df_flatfile_newrec.year <= year_max) #records to keep i_win = np.all(np.array([i_space_win_eq, i_space_win_sta, i_eq_depth, i_time_win]),axis=0) df_flatfile_newrec = df_flatfile_newrec[i_win] # Vs30 # ---- ---- ---- ---- ---- #Wills 2015 model for CA # Wills15Vs30 = pylib_W15_Vs30.Willis15Vs30CA() # df_flatfile_newrec.loc[:,'Vs30'] = Wills15Vs30.lookup(np.fliplr(df_flatfile_newrec[['staLat', 'staLon']]))[0] # df_flatfile_newrec.loc[df_flatfile_newrec.loc[:,'Vs30'] < 50,'Vs30'] = np.nan #geology based Vs30 if not df_sta_vs30 is None: #Vs30 estimates from geology and topography (Pengfei, personal communication) df_flatfile_newrec = pd.merge(df_flatfile_newrec, df_sta_vs30[['station','network','Vs30','Vs30sig']], on=['station','network'] ) else: #Unavailable Vs30 df_flatfile_newrec.loc[:,['Vs30','Vs30sig']] = np.nan # define earthquake, station and network ids # ---- ---- ---- ---- ---- #original earthquake and station ids df_flatfile_newrec.loc[:,'eventid'] = df_flatfile_newrec['eventid'].values.astype(int) df_flatfile_newrec.loc[:,'staid'] = np.unique(df_flatfile_newrec['station'], return_inverse=True)[1] + 1 df_flatfile_newrec.loc[:,'netid'] = np.unique(df_flatfile_newrec['network'], return_inverse=True)[1] + 1 #keep single record from each event i_unq_eq_sta = np.unique(df_flatfile_newrec[['eventid','staid','netid']].values, return_index=True, axis=0)[1] df_flatfile_newrec = df_flatfile_newrec.iloc[i_unq_eq_sta, :] # define rsn eqid and ssn # ---- ---- ---- ---- ---- #reset indices df_flatfile_newrec.reset_index(drop=True, inplace=True) df_flatfile_newrec.rename_axis('rsn', inplace=True) #updated earthquake and station ids _, eq_idx, eq_inv = np.unique(df_flatfile_newrec.loc[:,'eventid'], axis=0, return_index=True, return_inverse=True) _, sta_idx, sta_inv = np.unique(df_flatfile_newrec.loc[:,['netid','staid']], axis=0, return_index=True, return_inverse=True) df_flatfile_newrec.loc[:,'eqid'] = eq_inv+1 df_flatfile_newrec.loc[:,'ssn'] = sta_inv+1 n_eq_orig = len(eq_idx) n_sta_orig = len(sta_idx) # cartesian coordinates # ---- ---- ---- ---- ---- #utm coordinates eq_X = np.array([utmProj(e.eqLon, e.eqLat) for _, e in df_flatfile_newrec.iloc[eq_idx,:].iterrows()]) sta_X = np.array([utmProj(s.staLon, s.staLat) for _, s in df_flatfile_newrec.iloc[sta_idx,:].iterrows()]) df_flatfile_newrec.loc[:,['eqX','eqY']] = eq_X[eq_inv,:] df_flatfile_newrec.loc[:,['staX','staY']] = sta_X[sta_inv,:] #eq depth df_flatfile_newrec.loc[:,'eqZ'] = np.minimum(-1*df_flatfile_newrec['eqDepth'].values, 0) #utm zone df_flatfile_newrec.loc[:,'UTMzone'] = utm_zone # rupture distance # ---- ---- ---- ---- ---- df_flatfile_newrec.loc[:,'Rrup'] = np.sqrt(np.linalg.norm(df_flatfile_newrec[['eqX','eqY']].values-df_flatfile_newrec[['staX','staY']].values, axis=1)**2 + df_flatfile_newrec['eqZ']**2) #remove records based on rupture distance i_rrup = df_flatfile_newrec['Rrup'] < rrup_thres df_flatfile_newrec = df_flatfile_newrec.loc[i_rrup,:] # colocate stations # ---- ---- ---- ---- ---- #update ssn for colocated stations df_flatfile_newrec = pylib_catalog.ColocatePt(df_flatfile_newrec, 'ssn', ['staX','staY'], thres_dist=thres_dist) #keep single record from each event i_unq_eq_sta = np.unique(df_flatfile_newrec[['eqid','ssn']].values, return_index=True, axis=0)[1] df_flatfile_newrec = df_flatfile_newrec.iloc[i_unq_eq_sta, :].sort_index() # average gm parameters # ---- ---- ---- ---- ---- df_flatfile_newrec = pylib_catalog.IndexAvgColumns(df_flatfile_newrec, 'eqid', ['mag','eqX','eqY','eqZ']) df_flatfile_newrec = pylib_catalog.IndexAvgColumns(df_flatfile_newrec, 'ssn', ['Vs30','staX','staY','staElev']) #recalculated lat/lon coordinates _, eq_idx, eq_inv = np.unique(df_flatfile_newrec.loc[:,'eqid'], axis=0, return_index=True, return_inverse=True) _, sta_idx, sta_inv = np.unique(df_flatfile_newrec.loc[:,'ssn'], axis=0, return_index=True, return_inverse=True) n_eq = len(eq_idx) n_sta = len(sta_idx) eq_latlon = np.flip([utmProj(e.eqX, e.eqY, inverse=True) for _, e in df_flatfile_newrec.iloc[eq_idx,:].iterrows()], axis=1) sta_latlon = np.flip([utmProj(s.staX, s.staY, inverse=True) for _, s in df_flatfile_newrec.iloc[sta_idx,:].iterrows()], axis=1) df_flatfile_newrec.loc[:,['eqLat','eqLon']] = eq_latlon[eq_inv,:] df_flatfile_newrec.loc[:,['staLat','staLon']] = sta_latlon[sta_inv,:] # midpoint coordinates # ---- ---- ---- ---- ---- df_flatfile_newrec.loc[:,['mptX','mptY']] = (df_flatfile_newrec.loc[:,['eqX','eqY']].values + df_flatfile_newrec.loc[:,['staX','staY']].values) / 2 df_flatfile_newrec.loc[:,['mptLat','mptLon']] = np.flip( np.array([utmProj(pt.mptX, pt.mptY, inverse=True) for _, pt in df_flatfile_newrec.iterrows()]), axis=1 ) #recalculate rupture distance after averaging df_flatfile_newrec.loc[:,'Rrup'] = np.sqrt(np.linalg.norm(df_flatfile_newrec[['eqX','eqY']].values-df_flatfile_newrec[['staX','staY']].values, axis=1)**2 + df_flatfile_newrec['eqZ']**2) # %% Save Data # ====================================== # create output directories if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) if not os.path.isdir(dir_fig): pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) #full dataframe df_flatfile_full = df_flatfile_newrec[['eqid','ssn','eventid','staid','netid','station','network', 'mag','mag_type','mag_author','Rrup','Vs30','time','year', 'eqLat','eqLon','staLat','staLon','mptLat','mptLon', 'UTMzone','eqX','eqY','eqZ','staX','staY','staElev','mptX','mptY', 'author','cat','contributor','contributor_id','eq_loc']] #event dataframe df_flatfile_event = df_flatfile_newrec.iloc[eq_idx,:][['eqid','eventid','mag','mag_type','mag_author','year', 'eqLat','eqLon','UTMzone','eqX','eqY','eqZ', 'author','cat','contributor','contributor_id','eq_loc']].reset_index(drop=True) #station dataframe df_flatfile_station = df_flatfile_newrec.iloc[sta_idx,:][['ssn','Vs30', 'staLat','staLon','UTMzone','staX','staY','staElev']].reset_index(drop=True) # save dataframe # ---- ---- ---- ---- ---- #save processed dataframes fname_flatfile_full= '%s%s'%(dir_out, fname_flatfile) df_flatfile_full.to_csv(fname_flatfile_full + '.csv', index=True) df_flatfile_event.to_csv(fname_flatfile_full + '_event.csv', index=False) df_flatfile_station.to_csv(fname_flatfile_full + '_station.csv', index=False) # create figures # ---- ---- ---- ---- ---- # Mag-Dist distribution fname_fig = 'M-R_dist' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(df_flatfile_full.Rrup, df_flatfile_full.mag, label='new records') #edit figure properties ax.set_xlabel(r'Distance ($km$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') ax.set_xscale('log') # ax.set_xlim([0.1, 2000]) ax.set_ylim([1, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) # ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' ) # Mag-Year distribution fname_fig = 'M-date_dist' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(df_flatfile_event['year'].values, df_flatfile_event['mag'].values, label='new records') #edit figure properties ax.set_xlabel(r'time ($year$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') # ax.set_xscale('log') ax.set_xlim([1965, 2025]) ax.set_ylim([2, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) # ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' ) #eq and sta location fname_fig = 'eq_sta_locations' fig, ax, data_crs, gl = pylib_cplt.PlotMap(flag_grid=True) #plot earthquake and station locations ax.plot(df_flatfile_event['eqLon'].values, df_flatfile_event['eqLat'].values, '*', transform = data_crs, markersize = 10, zorder=12, label='Events') ax.plot(df_flatfile_station['staLon'].values, df_flatfile_station['staLat'].values, 'o', transform = data_crs, markersize = 6, zorder=13, label='Stations') #edit figure properties gl.xlabel_style = {'size': 25} gl.ylabel_style = {'size': 25} # gl.xlocator = mticker.FixedLocator([-124, -122, -120, -118, -116, -114]) # gl.ylocator = mticker.FixedLocator([32, 34, 36, 38, 40]) ax.legend(fontsize=25, loc='lower left') # ax.set_xlim(plt_latlon_win[:,1]) # ax.set_ylim(plt_latlon_win[:,0]) #save figure fig.tight_layout() fig.savefig( dir_fig + fname_fig + '.png' ) # print data info # ---- ---- ---- ---- ---- print(r'New Records:') print(f'\tnumber of rec: %.i'%len(df_flatfile_newrec)) print(f'\tnumber of rec (R<200km): %.i'%np.sum(df_flatfile_newrec.Rrup<=200)) print(f'\tnumber of rec (R<%.1f): %.i'%(rrup_thres, np.sum(df_flatfile_newrec.Rrup<=rrup_thres))) print(f'\tnumber of eq: %.i'%n_eq) print(f'\tnumber of sta: %.i'%n_sta) print(f'\tmin magnitude: %.1f'%df_flatfile_newrec.mag.min()) print(f'\tmax magnitude: %.1f'%df_flatfile_newrec.mag.max()) print(f'\tcoverage: %.i to %i'%(df_flatfile_newrec.year.min(), df_flatfile_newrec.year.max())) #write out summary # ---- ---- ---- ---- ---- f = open(dir_out + 'summary_data' + '.txt', 'w') f.write(f'New Records:\n') f.write(f'\tnumber of rec: %.i\n'%len(df_flatfile_newrec)) f.write(f'\tnumber of rec (R<200km): %.i\n'%np.sum(df_flatfile_newrec.Rrup<=200)) f.write(f'\tnumber of rec (R<%.1f): %.i\n'%(rrup_thres, np.sum(df_flatfile_newrec.Rrup<=rrup_thres))) f.write(f'\tnumber of eq: %.i\n'%n_eq) f.write(f'\tnumber of sta: %.i\n'%n_sta) f.write(f'\tmin magnitude: %.1f\n'%df_flatfile_newrec.mag.min()) f.write(f'\tmax magnitude: %.1f\n'%df_flatfile_newrec.mag.max()) f.write(f'\tcoverage: %.i to %i\n'%(df_flatfile_newrec.year.min(), df_flatfile_newrec.year.max())) f.close()
15,785
44.889535
162
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/preprocessing/PlotCellPaths.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Sep 13 18:00:32 2020 @author: glavrent """ # %% Required Packages # ====================================== #load variables import os import sys import pathlib import glob import re #regular expression package #arithmetic libraries import numpy as np import pandas as pd #geographic coordinates import pyproj #plottign libraries from matplotlib import pyplot as plt #user-derfined functions sys.path.insert(0,'../../Python_lib/plotting') import pylib_contour_plots as pycplt # Define Input Data # --------------------------- fname_flatfile = '../../../Data/Verification/preprocessing/flatfiles/CA_NV_2011-2021Lite/CatalogNewRecordsLite_2011-2021_CA_NV.csv' fname_cellinfo = '../../../Data/Verification/preprocessing/cell_distances/CatalogNGAWest3CALite_cellinfo.csv' fname_celldistfile = '../../../Data/Verification/preprocessing/cell_distances/CatalogNGAWest3CALite_distancematrix.csv' #grid limits and size coeff_latlon_win = np.array([[32, -125],[42.5, -114]]) #log scale for number of paths flag_logscl = True #output directory dir_out = '../../../Data/Verification/preprocessing/cell_distances/figures/' # Load Data # --------------------------- df_flatfile = pd.read_csv(fname_flatfile) df_cellinfo = pd.read_csv(fname_cellinfo) #cell distance file df_celldata = pd.read_csv(fname_celldistfile, index_col=0).reindex(df_flatfile.rsn) # Process Data # --------------------------- #coordinates and projection system # projection system utm_zone = np.unique(df_flatfile.UTMzone)[0] #utm zone utmProj = pyproj.Proj("+proj=utm +zone="+utm_zone+", +ellps=WGS84 +datum=WGS84 +units=m +no_defs") #cell edge coordinates cell_edge_latlon = [] for cell_edge in [['q5X','q5Y'], ['q6X','q6Y'], ['q8X','q8Y'], ['q7X','q7Y'], ['q5X','q5Y']]: cell_edge_latlon.append( np.fliplr(np.array([utmProj(c_xy[0]*1000, c_xy[1]*1000, inverse=True) for c_xy in df_cellinfo.loc[:,cell_edge].values])) ) cell_edge_latlon = np.hstack(cell_edge_latlon) #cell mid-coordinates cell_latlon = np.fliplr(np.array([utmProj(c_xy[0]*1000, c_xy[1]*1000, inverse=True) for c_xy in df_cellinfo.loc[:,['mptX','mptY']].values])) #earthquake and station ids eq_id_train = df_flatfile['eqid'].values.astype(int) sta_id_train = df_flatfile['ssn'].values.astype(int) eq_id, eq_idx_inv = np.unique(eq_id_train, return_index=True) sta_id, sta_idx_inv = np.unique(sta_id_train, return_index=True) #earthquake and station coordinates eq_latlon_train = df_flatfile[['eqLat', 'eqLon']].values stat_latlon_train = df_flatfile[['staLat', 'staLon']].values #unique earthquake and station coordinates eq_latlon = eq_latlon_train[eq_idx_inv,:] stat_latlon = stat_latlon_train[sta_idx_inv,:] #cell names cell_i = [bool(re.match('^c\\..*$',c_n)) for c_n in df_celldata.columns.values] #indices for cell columns cell_names = df_celldata.columns.values[cell_i] #cell-distance matrix with all cells cell_dist = df_celldata[cell_names] cell_n_paths = (cell_dist > 0).sum() # Create cell figures # --------------------------- if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) # if flag_pub: # # mpl.rcParams['font.family'] = 'Avenir' # plt.rcParams['axes.linewidth'] = 2 # Plot cell paths fname_fig = 'cA_paths' fig, ax, data_crs, gl = pycplt.PlotMap() #plot earthquake and station locations ax.plot(eq_latlon[:,1], eq_latlon[:,0], '*', transform = data_crs, markersize = 10, zorder=13, label='Events') ax.plot(stat_latlon[:,1], stat_latlon[:,0], 'o', transform = data_crs, markersize = 6, zorder=12, label='Stations') # ax.plot(eq_latlon[:,1], eq_latlon[:,0], '^', transform = data_crs, color = 'black', markersize = 10, zorder=13, label='Earthquake') # ax.plot(stat_latlon[:,1], stat_latlon[:,0], 'o', transform = data_crs, color = 'black', markersize = 3, zorder=12, label='Station') #plot earthquake-station paths for rec in df_flatfile[['eqLat','eqLon','staLat','staLon']].iterrows(): ax.plot(rec[1][['eqLon','staLon']], rec[1][['eqLat','staLat']], transform = data_crs, color = 'gray', linewidth=0.05, zorder=10, alpha=0.2) #plot cells for ce_xy in cell_edge_latlon: ax.plot(ce_xy[[1,3,5,7,9]],ce_xy[[0,2,4,6,8]],color='gray', transform = data_crs) #figure limits ax.set_xlim( coeff_latlon_win[:,1] ) ax.set_ylim( coeff_latlon_win[:,0] ) #edit figure properties #grid lines gl = ax.gridlines(draw_labels=True) gl.xlabels_top = False gl.ylabels_right = False gl.xlabel_style = {'size': 25} gl.ylabel_style = {'size': 25} #add legend ax.legend(fontsize=25, loc='lower left') #apply tight layout fig.show() fig.tight_layout() fig.savefig( dir_out + fname_fig + '.png') # Plot cell paths fname_fig = 'cA_num_paths' cbar_label = 'Number of paths' data2plot = np.vstack([cell_latlon.T, cell_n_paths.values]).T #color limits cmin = 0 cmax = 2000 #log scale options if flag_logscl: # data2plot[:,2] = np.maximum(data2plot[:,2], 1) cmin = np.log(1) cmax = np.log(cmax) #create figure fig, ax, cbar, data_crs, gl = pycplt.PlotCellsCAMap(data2plot, cmin=cmin, cmax=cmax, log_cbar = flag_logscl, frmt_clb = '%.0f', cmap='OrRd') #plot cells for ce_xy in cell_edge_latlon: ax.plot(ce_xy[[1,3,5,7]],ce_xy[[0,2,4,6]],color='gray', transform = data_crs) #figure limits ax.set_xlim( coeff_latlon_win[:,1] ) ax.set_ylim( coeff_latlon_win[:,0] ) #edit figure properties #grid lines gl = ax.gridlines(draw_labels=True) gl.xlabels_top = False gl.ylabels_right = False gl.xlabel_style = {'size': 25} gl.ylabel_style = {'size': 25} #update colorbar cbar.set_label(cbar_label, size=30) cbar.ax.tick_params(labelsize=25) #apply tight layout fig.show() fig.tight_layout() fig.savefig( dir_out + fname_fig + '.png')
5,935
34.54491
143
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/preprocessing/PlotUsableMagRrupCatalog.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Mon Oct 4 16:32:37 2021 @author: glavrent """ # %% Required Packages # ====================================== #load libraries import os import pathlib #arithmetic libraries import numpy as np import pandas as pd #plotting libraries from matplotlib import pyplot as plt import matplotlib.ticker as mticker # %% Define variables # ====================================== #input file names fname_flatfile_NGA2 = '../../../Raw_files/nga_w2/Updated_NGA_West2_Flatfile_RotD50_d050_public_version.xlsx' fname_mag_rrup_lim = '../../../Data/Verification/preprocessing/flatfiles/usable_mag_rrup/usable_Mag_Rrup_coeffs.csv' #output directoy dir_fig = '../../../Data/Verification/preprocessing/flatfiles/usable_mag_rrup/' # %% Load Data # ====================================== #NGAWest2 df_flatfile_NGA2 = pd.read_excel(fname_flatfile_NGA2) #M/R limit df_m_r_lim = pd.read_csv(fname_mag_rrup_lim,index_col=0) #remove rec with unavailable data df_flatfile_NGA2 = df_flatfile_NGA2.loc[df_flatfile_NGA2.EQID>0,:] df_flatfile_NGA2 = df_flatfile_NGA2.loc[df_flatfile_NGA2['ClstD (km)']>0,:] #mag and distance arrays mag_array = df_flatfile_NGA2['Earthquake Magnitude'] rrup_array = df_flatfile_NGA2['ClstD (km)'] #compute limit rrup_lim1 = np.arange(0,1001) mag_lim1 = (df_m_r_lim.loc['b0','coefficients'] + df_m_r_lim.loc['b1','coefficients'] * rrup_lim1 + df_m_r_lim.loc['b2','coefficients'] * rrup_lim1**2) rrup_lim2 = df_m_r_lim.loc['max_rrup','coefficients'] # %% Process Data # ====================================== if not os.path.isdir(dir_fig): pathlib.Path(dir_fig).mkdir(parents=True, exist_ok=True) # create figures # ---- ---- ---- ---- ---- # Mag-Dist distribution fname_fig = 'M-R_limits' #create figure fig, ax = plt.subplots(figsize = (10,9)) pl1 = ax.scatter(rrup_array, mag_array, label='NGAWest2 CA') pl2 = ax.plot(rrup_lim1, mag_lim1, linewidth=2, color='black') pl3 = ax.vlines(rrup_lim2, ymin=0, ymax=10, linewidth=2, color='black', linestyle='--') #edit figure properties ax.set_xlabel(r'Distance ($km$)', fontsize=30) ax.set_ylabel(r'Magnitude', fontsize=30) ax.grid(which='both') # ax.set_xscale('log') ax.set_xlim([0, 1000]) ax.set_ylim([2, 8]) ax.tick_params(axis='x', labelsize=25) ax.tick_params(axis='y', labelsize=25) # ax.legend(fontsize=25, loc='upper left') ax.xaxis.set_tick_params(which='major', size=10, width=2, direction='in', top='on') ax.xaxis.set_tick_params(which='minor', size=7, width=2, direction='in', top='on') ax.yaxis.set_tick_params(which='major', size=10, width=2, direction='in', right='on') ax.yaxis.set_tick_params(which='minor', size=7, width=2, direction='in', right='on') fig.tight_layout() #save figure fig.savefig( dir_fig + fname_fig + '.png' )
2,814
33.753086
117
py
ngmm_tools
ngmm_tools-master/Analyses/Code_Verification/preprocessing/ComputeCellDistance.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Apr 9 11:04:25 2020 @author: glavrent """ # Required Packages # ====================================== #load libraries import os import sys import pathlib import numpy as np import pandas as pd from scipy import sparse #geographic libraries import pyproj #user libraries sys.path.insert(0,'../../Python_lib/ground_motions') import pylib_cell_dist as pylib_cells # %% Define Input Data # ====================================== #input flatfile # fname_flatfile = 'CatalogNGAWest3CA' # fname_flatfile = 'CatalogNGAWest3CA_2013' fname_flatfile = 'CatalogNGAWest3CALite' # fname_flatfile = 'CatalogNGAWest3NCA' # fname_flatfile = 'CatalogNGAWest3SCA' dir_flatfile = '../../../Data/Verification/preprocessing/flatfiles/merged/' #output files dir_out = '../../../Data/Verification/preprocessing/cell_distances/' # %% Read and Porcess Input Data # ====================================== # read ground-motion data fullname_flatfile = dir_flatfile + fname_flatfile + '.csv' df_flatfile = pd.read_csv(fullname_flatfile) n_rec = len(df_flatfile) #define projection system assert(len(np.unique(df_flatfile.UTMzone))==1),'Error. Multiple UTM zones defined.' utm_zone = df_flatfile.UTMzone[0] utmProj = pyproj.Proj("+proj=utm +zone="+utm_zone+", +ellps=WGS84 +datum=WGS84 +units=m +no_defs") #create output directory if not os.path.isdir(dir_out): pathlib.Path(dir_out).mkdir(parents=True, exist_ok=True) #create object with source and station locations #utm coordinates data4celldist = df_flatfile.loc[:,['eqX','eqY','eqZ','staX','staY']].values flagUTM = True #add elevation for stations data4celldist = np.hstack([data4celldist,np.zeros([n_rec,1])]) # %% Create Cell Grid # ====================================== #grid range grid_lims_x = [data4celldist[:,[0,3]].min(), data4celldist[:,[0,3]].max()] grid_lims_y = [data4celldist[:,[1,4]].min(), data4celldist[:,[1,4]].max()] grid_lims_z = [data4celldist[:,[2,5]].min(), data4celldist[:,[2,5]].max()] #manual limits #utm limits # #NGAWest3 full # grid_lims_x = [-200, 1100] # grid_lims_y = [3300, 4800] # grid_lims_z = [-100, 0] #NGAWest3 lite grid_lims_x = [-200, 800] grid_lims_y = [3450, 4725] grid_lims_z = [-50, 0] #cell size cell_size = [25, 25, 50] #lat-lon grid spacing grid_x = np.arange(grid_lims_x[0], grid_lims_x[1]+0.1, cell_size[0]) grid_y = np.arange(grid_lims_y[0], grid_lims_y[1]+0.1, cell_size[1]) grid_z = np.arange(grid_lims_z[0], grid_lims_z[1]+0.1, cell_size[2]) #cell schematic # (7)-----(8) (surface, top face) # / | / | # (5)-----(6) | # | | | | # | (3)---|-(4) (bottom face) # |/ |/ # (1)-----(2) #create cells j1 = 0 j2 = 0 j3 = 0 cells = [] for j1 in range(len(grid_x)-1): for j2 in range(len(grid_y)-1): for j3 in range(len(grid_z)-1): #cell corners (bottom-face) cell_c1 = [grid_x[j1], grid_y[j2], grid_z[j3]] cell_c2 = [grid_x[j1+1], grid_y[j2], grid_z[j3]] cell_c3 = [grid_x[j1], grid_y[j2+1], grid_z[j3]] cell_c4 = [grid_x[j1+1], grid_y[j2+1], grid_z[j3]] #cell corners (top-face) cell_c5 = [grid_x[j1], grid_y[j2], grid_z[j3+1]] cell_c6 = [grid_x[j1+1], grid_y[j2], grid_z[j3+1]] cell_c7 = [grid_x[j1], grid_y[j2+1], grid_z[j3+1]] cell_c8 = [grid_x[j1+1], grid_y[j2+1], grid_z[j3+1]] #cell center cell_cent = np.mean(np.stack([cell_c1,cell_c2,cell_c3,cell_c4, cell_c5,cell_c6,cell_c7,cell_c8]),axis = 0).tolist() #summarize all cell coordinates in a list cell_info = cell_c1 + cell_c2 + cell_c3 + cell_c4 + \ cell_c5 + cell_c6 + cell_c7 + cell_c8 + cell_cent #add cell info cells.append(cell_info) del j1, j2, j3, cell_info del cell_c1, cell_c2, cell_c3, cell_c4, cell_c5, cell_c6, cell_c7, cell_c8 cells = np.array(cells) n_cells = len(cells) #cell info cell_ids = np.arange(n_cells) cell_names = ['c.%i'%(i) for i in cell_ids] cell_q_names = ['q1X','q1Y','q1Z','q2X','q2Y','q2Z','q3X','q3Y','q3Z','q4X','q4Y','q4Z', 'q5X','q5Y','q5Z','q6X','q6Y','q6Z','q7X','q7Y','q7Z','q8X','q8Y','q8Z', 'mptX','mptY','mptZ'] # Create cell info dataframe # ---- ---- ---- ---- ---- #cell names df_data1 = pd.DataFrame({'cellid': cell_ids, 'cellname': cell_names}) #cell coordinates df_data2 = pd.DataFrame(cells, columns = cell_q_names) df_cellinfo = pd.merge(df_data1,df_data2,left_index=True,right_index=True) # add cell utm zone df_cellinfo.loc[:,'UTMzone'] = utm_zone # Compute Lat\Lon of cells # ---- ---- ---- ---- ---- #cell verticies for q in range(1,9): c_X = ['q%iX'%q, 'q%iY'%q] c_latlon = ['q%iLat'%q, 'q%iLon'%q] df_cellinfo.loc[:,c_latlon] = np.flip( np.array([utmProj(pt_xy[0]*1e3, pt_xy[1]*1e3, inverse=True) for _, pt_xy in df_cellinfo[c_X].iterrows() ]), axis=1) #cell midpoints c_X = ['mptX', 'mptY'] c_latlon = ['mptLat','mptLon'] df_cellinfo.loc[:,c_latlon] = np.flip( np.array([utmProj(pt_xy[0]*1e3, pt_xy[1]*1e3, inverse=True) for _, pt_xy in df_cellinfo[c_X].iterrows() ]), axis=1) # %% Compute Cell distances # ====================================== cells4dist = cells[:,[0,1,2,21,22,23]] distancematrix = np.zeros([len(data4celldist), len(cells4dist)]) for i in range(len(data4celldist)): print('Computing cell distances, record',i) pt1 = data4celldist[i,(0,1,2)] pt2 = data4celldist[i,(3,4,5)] dm = pylib_cells.ComputeDistGridCells(pt1,pt2,cells4dist, flagUTM) distancematrix[i] = dm #print Rrup missfits dist_diff = df_flatfile.Rrup - distancematrix.sum(axis=1) print('max R_rup misfit', max(dist_diff)) print('min R_rup misfit', min(dist_diff)) #convert cell distances to sparse matrix distmatrix_sparce = sparse.coo_matrix(distancematrix) # Create cell distances data-frame # ---- ---- ---- ---- ---- #record info df_recinfo = df_flatfile[['rsn','eqid','ssn']] #cell distances df_celldist = pd.DataFrame(distancematrix, columns = cell_names) df_celldist = pd.merge(df_recinfo, df_celldist, left_index=True, right_index=True) #spase cell distances df_celldist_sp = pd.DataFrame({'row': distmatrix_sparce.row+1, 'col': distmatrix_sparce.col+1, 'data': distmatrix_sparce.data}) # %% Save data # ====================================== #save cell info fname_cellinfo = fname_flatfile + '_cellinfo' df_cellinfo.to_csv(dir_out + fname_cellinfo + '.csv', index=False) # #save distance metrics fname_celldist = fname_flatfile + '_distancematrix' df_celldist.to_csv(dir_out + fname_celldist + '.csv', index=False) # #save distance matrix as sparce fname_celldist = fname_flatfile + '_distancematrix_sparce' df_celldist_sp.to_csv(dir_out + fname_celldist + '.csv', index=False)
7,014
33.219512
127
py