jpxkqx commited on
Commit
0a1702b
1 Parent(s): ca24e7a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -8
README.md CHANGED
@@ -168,8 +168,8 @@ rectification, and grid interpolation. The methodology employed in each step is
168
 
169
  - Inference speed: The ConvSwin2SR model demonstrates a commendable inference speed, particularly when handling a substantial batch of samples.
170
  Specifically, when tasked with downscaling 248 samples, which is synonymous with processing data for an entire month at 3-hour intervals,
171
- the model completes the operation in a mere 21 seconds. This level of efficiency is observed in a local computing environment outfitted with 16GB o
172
- f RAM and 4GB of GPU memory.
173
 
174
  # Evaluation
175
 
@@ -233,17 +233,23 @@ The Swin2 transformer optimizes its parameters using a composite loss function t
233
  accuracy across different resolutions and representations:
234
 
235
  1. **Primary Predictions Loss**:
236
- - This term computes the L1 loss between the primary model predictions and the reference values. It ensures that the transformer's outputs
237
- closely match the ground truth across the primary spatial resolution.
238
 
239
  2. **Downsampled Predictions Loss**:
240
- - Recognizing the importance of accuracy across varying resolutions, this term calculates the L1 loss between the downsampled versions of the
241
- predictions and the reference values. By incorporating this term, the model is incentivized to preserve critical information even when the data is represented
242
- at a coarser scale.
 
 
243
 
244
  3. **Blurred Predictions Loss**:
245
  - To ensure the model's robustness against small perturbations and noise, this term evaluates the L1 loss between blurred versions of the
246
- predictions and the references. This encourages the model to produce predictions that maintain accuracy even under slight modifications in the data representation.
 
 
 
 
247
 
248
  ## Computing Infrastructure
249
 
 
168
 
169
  - Inference speed: The ConvSwin2SR model demonstrates a commendable inference speed, particularly when handling a substantial batch of samples.
170
  Specifically, when tasked with downscaling 248 samples, which is synonymous with processing data for an entire month at 3-hour intervals,
171
+ the model completes the operation in a mere 21 seconds. This level of efficiency is observed in a local computing environment outfitted with 16GB of
172
+ RAM and 4GB of GPU memory.
173
 
174
  # Evaluation
175
 
 
233
  accuracy across different resolutions and representations:
234
 
235
  1. **Primary Predictions Loss**:
236
+ - This term computes the L1 loss between the primary model predictions and the reference values. It ensures that the transformer's
237
+ outputs closely match the ground truth.
238
 
239
  2. **Downsampled Predictions Loss**:
240
+ - This term calculates the L1 loss between the downsampled versions of the predictions and the reference values. By incorporating this term,
241
+ the model is incentivized to preserve the underlying relations between both spatial resolutions. The references and predictions are upscaled
242
+ by average pooling by a factor of x5 to match the source resolution. Although this loss term could be (technically) computed with respect
243
+ to the low-resolution sample, the upscaled reference values are considered, due to the fact that the average pooling used for upscaling does
244
+ not represent the true relationship between both datasets considered.
245
 
246
  3. **Blurred Predictions Loss**:
247
  - To ensure the model's robustness against small perturbations and noise, this term evaluates the L1 loss between blurred versions of the
248
+ predictions and the references. This encourages the model to produce predictions that maintain accuracy even under slight modifications
249
+ in the data representation. On the other hand, it can smooth the prediction field too much, so it is a term whose use should be studied
250
+ before including it in your model. To produce the blurred values, a gaussian kernel of size 5 is applied.
251
+
252
+ By combining these loss terms, the ConvSwin2SR is trained to produce realistic predictions.
253
 
254
  ## Computing Infrastructure
255