_id
stringlengths 36
36
| text
stringlengths 5
665k
| marker
stringlengths 3
6
| marker_offsets
sequence | label
stringlengths 28
32
|
---|---|---|---|---|
f0fac466-5ba7-4854-9774-102235d25de5 | After the deviation of the appropriate GS-TWDP statistical expressions, it is also necessary to verify the applicability of proposed distribution for modeling propagation in mmWave bands.
In that sense, GS-TWDP model is fitted to a measurement results performed at 28 GHz and published in [1]}. Obtained results are then expressed in terms of a fitting error calculated using modified version of the Kolmogorov-Smirnov (KS) [2]}, in order to outweigh the fit in amplitude values closer to zero, for which the fading is more severe [2]}.
Accordingly, goodness of fit between empirical and theoretical CDFs, denoted by \(\hat{F}_{r_c}(x)\) and \(F_{r_c}(x)\) respectively, is expressed as [2]}:
\(\epsilon = \max \left|\log _{10}\hat{F}_r(x)-\log _{10}F_{r_c}(x)\right|\)
| [2] | [
[
424,
427
],
[
531,
534
],
[
689,
692
]
] | https://openalex.org/W3098595577 |
e20733bf-804a-41be-8126-65c9b0a600f9 | Heart Rate Variability (HRV) [1]}, [2]} is a reliable measure used in physiological and psychological research, since HRV reflects cardiac autonomic nervous system (ANS) regulation variation (i.e., the changing balance between the sympathetic and parasympathetic nervous systems).
Beyond a purely cardiovascular measure, HRV has been used as a tool to detect acute illness (e.g., early detection of COVID-19 [3]}), common cold [4]}, and inflammatory response from infection [5]}).
HRV also aids diagnostics [1]}, and is predictive of various cardiovascular-related diseases [7]}.
Numerous studies have investigated the connection between HRV and a range of conditions, including cardiovascular [8]}, [9]}, [10]}, blood pressure [11]}, [12]}, and myocardial infarction [13]}, [14]}, [15]}.
| [1] | [
[
29,
32
],
[
507,
510
]
] | https://openalex.org/W2026699230 |
4d5b2a49-a1c3-4ab2-af91-2288e1dee2ec | Note that joint estimation of the coupling and map, as equation (REF ) does, is shown to be effective in a few OT studies [1]}, [2]}, but this is the first time optimal transport map estimation is used in a multi-task regression.
| [1] | [
[
122,
125
]
] | https://openalex.org/W2963237040 |
86b00c84-b217-4bd2-a134-b84ea79194aa | When fixing \(\mathbf {W}\) , the second term of equation (REF ) is related to a category of OT formulation that can be solved by alternatively minimizing [1]} over coupling \(\pi \) and transformation \(\mathbf {F}\) .
Specifically, when solving for \(\pi \) with \(\mathbf {W}\) and \(\mathbf {F}\) fixed, the objective of equation (REF ) becomes an \(O(n^3 \log n)\) linear program.
Here, adding an entropy regularization will allow us to use the celebrated Sinkhorn algorithm Ag.(REF ) [2]} that just requires \(O(n^2)\) operations [3]}, as:
\(\pi ^* = \arg \min _{\pi \in \Pi } \sum _{i,j=1}^T \pi _{i,j} {\mathbf {C}}_{i,j} + \gamma H(\pi ),\)
| [1] | [
[
155,
158
]
] | https://openalex.org/W2963237040 |
c3cac511-7df7-429e-b03a-bee14aaa56f5 | To tackle domain adaptation limitations, we incorporate optimal transport map estimation theory with multi-task learning. Optimal transport (OT) [1]} has attracted considerable interest in the machine learning community.
OT provides a meaningful distance between probability measures that do not share the same support; OT also formulates a transport map [2]}, [3]}, [4]} that pushes forward one distribution onto another. Also, the addition of entropic regularization [5]} yields fast computation, thus enabling scalability of OT to larger problems.
| [4] | [
[
367,
370
]
] | https://openalex.org/W2963237040 |
cc10035e-587e-4381-8fd4-4415a0fa9efd | Remote sensing (RS) ship detection has attracted extensive attention in recent years due to its large potential in both civilian and military applications (e.g., port management, target surveillance). As a key factor of ship detection, high-resolution (HR) images can provide abundant appearance information and thus introduce improvement to the detection accuracy [1]}. However, obtaining an HR image posts a high requirement on the satellite sensors and generally results in an expensive cost. Consequently, using image super-resolution (SR) techniques to recover the missing details in RS images has become a popular research topic and has been widely investigated in recent years.
| [1] | [
[
365,
368
]
] | https://openalex.org/W2963606888 |
84c69712-176d-45a4-922d-67df64963b37 | Aiming at the aforementioned issue, in this paper, we propose an end-to-end network named ShipSRDet to fully use the super-resolved feature representation for RS ship detection. Different from existing SR-based detection methods, we not only feed the super-resolved images to the detector, but also integrate the intermediate features of the SR network with those of the detection network. In this way, more informative cues can be transferred from the SR network to the detection network to enhance the detection performance. Experiments on the HRSC dataset [1]} demonstrate the effectiveness of our method. As shown in Fig. REF , our ShipSRDet achieves notable improvements on detection performance, and recovers the missing details in the input image.
<FIGURE> | [1] | [
[
559,
562
]
] | https://openalex.org/W2594177559 |
9f848584-569e-468f-b530-ff6f5c10a6c6 | As shown in Fig. REF (a), our SR module takes a medium-low resolution image \(\mathcal {I}_{LR}\in \mathbb {R}^{H\times W\times 3}\) as its input to produce an SR image \(\mathcal {I}_{SR}\in \mathbb {R}^{\alpha H\times \alpha W\times 3}\) and two intermediate features \(\mathcal {F}^{LR}_{out}\in \mathbb {R}^{H\times W\times 64}\) and \(\mathcal {F}^{HR}_{out}\in \mathbb {R}^{\alpha H\times \alpha W\times 64}\) , where \(H\) and \(W\) represent the height and width of the input image, and \(\alpha \) denotes the upscaling factor. Here, we use the residual dense block (RDB) [1]} as the basic block in our SR module since it can fully use features from all preceding layers to generate hierarchical representations, which is demonstrated beneficial to SR reconstruction [2]}.
| [1] | [
[
587,
590
]
] | https://openalex.org/W2964101377 |
b692f859-80b4-4398-b582-35e96f5cbba5 | As shown in Fig. REF (a), our SR module takes a medium-low resolution image \(\mathcal {I}_{LR}\in \mathbb {R}^{H\times W\times 3}\) as its input to produce an SR image \(\mathcal {I}_{SR}\in \mathbb {R}^{\alpha H\times \alpha W\times 3}\) and two intermediate features \(\mathcal {F}^{LR}_{out}\in \mathbb {R}^{H\times W\times 64}\) and \(\mathcal {F}^{HR}_{out}\in \mathbb {R}^{\alpha H\times \alpha W\times 64}\) , where \(H\) and \(W\) represent the height and width of the input image, and \(\alpha \) denotes the upscaling factor. Here, we use the residual dense block (RDB) [1]} as the basic block in our SR module since it can fully use features from all preceding layers to generate hierarchical representations, which is demonstrated beneficial to SR reconstruction [2]}.
| [2] | [
[
782,
785
]
] | https://openalex.org/W3175560234 |
2e816274-aee1-4030-a298-d676cc3ccb73 | Specifically, the input image \(\mathcal {I}_{LR}\) is first fed to a \(3\times 3\) convolution to generate initial feature \(\mathcal {F}_0 \in \mathbb {R}^{H\times W\times 64}\) . Then, \(\mathcal {F}_0\) is fed to 8 cascaded RDBs for deep feature extraction. Within each RDB, we use 5 convolutions with a growth rate of 32. As shown in Fig. REF (b), features from all the layers in an RDB are concatenated and fed to a \(1\times 1\) convolution for local fusion. Similarly, features from all the RDBs in our SR module are concatenated for global fusion. Afterwards, the fused feature \(\mathcal {F}^{LR}_{out}\) is added with the initial feature \(\mathcal {F}_0\) and fed to a sub-pixel layer [1]} to generate the upsampled feature \(\mathcal {F}^{HR}_{out}\) . Finally, \(\mathcal {F}^{HR}_{out}\) is fed to a \(3\times 3\) convolution to produce the residual prediction which is further added with the bicubicly upsampled input image to generate the final SR image \(\mathcal {I}_{SR}\) .
| [1] | [
[
703,
706
]
] | https://openalex.org/W2476548250 |
78d34211-ad27-4864-aa59-857bffb2de07 | Single shot multi-box detector (SSD) [1]} is used as the detection module in our ShipSRDet. In our detection module, VGG-16 network [2]} is used to extract multi-scale features from the input SR image \(\mathcal {I}_{SR}\) . Simultaneously, feature adaption is performed to integrate features from the SR module with those extracted by the VGG network. Specifically, two transition convolutions are performed on \(\mathcal {F}^{LR}_{out}\) and \(\mathcal {F}^{HR}_{out}\) to adjust their feature depth to 256 and 64 to produce feature \(\mathcal {F}^{LR}_{SSD}\) and \(\mathcal {F}^{HR}_{SSD}\) , respectively. Then, \(\mathcal {F}^{HR}_{SSD}\) and \(\mathcal {F}^{LR}_{SSD}\) are added to the features in the 2nd and 7th layer of the VGG network to achieve feature integration. After initial feature extraction with the VGG network, an encoder is further employed for high-level feature extraction. Finally, seven feature maps of different resolutions are generated and fed into the detector\(\&\) classifier to predict the location and category candidates. The final prediction is produced by performing fast non-maximum suppression (fast NMS) on the candidates.
| [1] | [
[
37,
40
]
] | https://openalex.org/W3106250896 |
1c122b70-874c-4a7f-8db3-045ffd43c376 | Single shot multi-box detector (SSD) [1]} is used as the detection module in our ShipSRDet. In our detection module, VGG-16 network [2]} is used to extract multi-scale features from the input SR image \(\mathcal {I}_{SR}\) . Simultaneously, feature adaption is performed to integrate features from the SR module with those extracted by the VGG network. Specifically, two transition convolutions are performed on \(\mathcal {F}^{LR}_{out}\) and \(\mathcal {F}^{HR}_{out}\) to adjust their feature depth to 256 and 64 to produce feature \(\mathcal {F}^{LR}_{SSD}\) and \(\mathcal {F}^{HR}_{SSD}\) , respectively. Then, \(\mathcal {F}^{HR}_{SSD}\) and \(\mathcal {F}^{LR}_{SSD}\) are added to the features in the 2nd and 7th layer of the VGG network to achieve feature integration. After initial feature extraction with the VGG network, an encoder is further employed for high-level feature extraction. Finally, seven feature maps of different resolutions are generated and fed into the detector\(\&\) classifier to predict the location and category candidates. The final prediction is produced by performing fast non-maximum suppression (fast NMS) on the candidates.
| [2] | [
[
132,
135
]
] | https://openalex.org/W1686810756 |
13d5b6d0-eed6-4da9-a8ca-7233500e5b05 | We used the HRSC dataset [1]} for both training and test. We followed the original dataset split and performed \(2\times \) and \(8\times \) bicubic downsampling to generate training and test image pairs. Consequently, each HR image has a resolution of \(512\times 512\) and its corresponding medium-low resolution conterpart has a resolution of \(128\times 128\) . Random horizontal flipping, random vertical flipping and random rotation were performed for data augmentation.
| [1] | [
[
25,
28
]
] | https://openalex.org/W2594177559 |
df6b961f-0994-448b-9ef9-6a7c5bd6f53b | Our ShipSRDet was implemented in PyTorch on a PC with an RTX 2080Ti GPU, and trained in a two-stage pipeline. In the first stage, we followed [1]}, [2]} to train our SR module using the generated image pairs with an \(L_1\) loss. Adam method [3]} is used for optimization. The batch size was set to 4 and the learning rate was was initially set to \(1\times 10^{-4}\) and halved for every 200 epochs. The training was stopped after 450 epochs. In the second stage, we concatenated the SR module with the pre-trained detection moduleWe used the publicly available SSD network which was pre-trained on the COCO [4]} dataset., and performed end-to-end finetuning for global optimization. In the finetuning stage, the learning rate was initially set to \(1\times 10^{-4}\) and decreased by a factor of 0.1 for every 10 epochs. The finetuning process was performed for 24 epochs.
| [1] | [
[
142,
145
]
] | https://openalex.org/W3110151525 |
f4006391-4fa6-49c9-9051-846aae464bb7 | Our ShipSRDet was implemented in PyTorch on a PC with an RTX 2080Ti GPU, and trained in a two-stage pipeline. In the first stage, we followed [1]}, [2]} to train our SR module using the generated image pairs with an \(L_1\) loss. Adam method [3]} is used for optimization. The batch size was set to 4 and the learning rate was was initially set to \(1\times 10^{-4}\) and halved for every 200 epochs. The training was stopped after 450 epochs. In the second stage, we concatenated the SR module with the pre-trained detection moduleWe used the publicly available SSD network which was pre-trained on the COCO [4]} dataset., and performed end-to-end finetuning for global optimization. In the finetuning stage, the learning rate was initially set to \(1\times 10^{-4}\) and decreased by a factor of 0.1 for every 10 epochs. The finetuning process was performed for 24 epochs.
| [2] | [
[
148,
151
]
] | https://openalex.org/W3111051040 |
009cc96c-ac28-42f5-8382-4c3885a04194 | Our ShipSRDet was implemented in PyTorch on a PC with an RTX 2080Ti GPU, and trained in a two-stage pipeline. In the first stage, we followed [1]}, [2]} to train our SR module using the generated image pairs with an \(L_1\) loss. Adam method [3]} is used for optimization. The batch size was set to 4 and the learning rate was was initially set to \(1\times 10^{-4}\) and halved for every 200 epochs. The training was stopped after 450 epochs. In the second stage, we concatenated the SR module with the pre-trained detection moduleWe used the publicly available SSD network which was pre-trained on the COCO [4]} dataset., and performed end-to-end finetuning for global optimization. In the finetuning stage, the learning rate was initially set to \(1\times 10^{-4}\) and decreased by a factor of 0.1 for every 10 epochs. The finetuning process was performed for 24 epochs.
| [3] | [
[
243,
246
]
] | https://openalex.org/W2964121744 |
b4f1e715-40e3-4a17-8251-919eb1eca7a5 | Our ShipSRDet was implemented in PyTorch on a PC with an RTX 2080Ti GPU, and trained in a two-stage pipeline. In the first stage, we followed [1]}, [2]} to train our SR module using the generated image pairs with an \(L_1\) loss. Adam method [3]} is used for optimization. The batch size was set to 4 and the learning rate was was initially set to \(1\times 10^{-4}\) and halved for every 200 epochs. The training was stopped after 450 epochs. In the second stage, we concatenated the SR module with the pre-trained detection moduleWe used the publicly available SSD network which was pre-trained on the COCO [4]} dataset., and performed end-to-end finetuning for global optimization. In the finetuning stage, the learning rate was initially set to \(1\times 10^{-4}\) and decreased by a factor of 0.1 for every 10 epochs. The finetuning process was performed for 24 epochs.
| [4] | [
[
611,
614
]
] | https://openalex.org/W1861492603 |
5dbf03f1-8c74-4586-9f7f-4eba6e8c4a56 | Discontinuous Galerkin (dG) finite element methods were introduced in the early 1970s for the numerical solution of first-order hyperbolic problems [1]} and the weak imposition of inhomogeneous boundary conditions for elliptic problems [2]}. In the past several decades, dG methods have enjoyed considerable success as a standard variational framework for the numerical solution of many classes of problems involving partial differential equations (PDEs); see, e.g., monographs [3]}, [4]}, [5]} for reviews of some of the main developments of dG methods.
The interest in dG methods can be attributed to a number of factors, including the great flexibility in dealing with \(hp\) -adaptivity and general shaped elements [6]}, [7]}, [8]},
as well as in solving convection-dominated PDEs; see, e.g., early works [9]}, [10]} concerning hyperbolic conservation laws and convection-diffusion problems.
| [1] | [
[
148,
151
]
] | https://openalex.org/W49087508 |
c9f5de7e-3bd7-4a4e-8a28-cbab896746de | Discontinuous Galerkin (dG) finite element methods were introduced in the early 1970s for the numerical solution of first-order hyperbolic problems [1]} and the weak imposition of inhomogeneous boundary conditions for elliptic problems [2]}. In the past several decades, dG methods have enjoyed considerable success as a standard variational framework for the numerical solution of many classes of problems involving partial differential equations (PDEs); see, e.g., monographs [3]}, [4]}, [5]} for reviews of some of the main developments of dG methods.
The interest in dG methods can be attributed to a number of factors, including the great flexibility in dealing with \(hp\) -adaptivity and general shaped elements [6]}, [7]}, [8]},
as well as in solving convection-dominated PDEs; see, e.g., early works [9]}, [10]} concerning hyperbolic conservation laws and convection-diffusion problems.
| [2] | [
[
236,
239
]
] | https://openalex.org/W2079423956 |
7f766b4a-6a4c-45ca-af74-13a8ce83f059 | Discontinuous Galerkin (dG) finite element methods were introduced in the early 1970s for the numerical solution of first-order hyperbolic problems [1]} and the weak imposition of inhomogeneous boundary conditions for elliptic problems [2]}. In the past several decades, dG methods have enjoyed considerable success as a standard variational framework for the numerical solution of many classes of problems involving partial differential equations (PDEs); see, e.g., monographs [3]}, [4]}, [5]} for reviews of some of the main developments of dG methods.
The interest in dG methods can be attributed to a number of factors, including the great flexibility in dealing with \(hp\) -adaptivity and general shaped elements [6]}, [7]}, [8]},
as well as in solving convection-dominated PDEs; see, e.g., early works [9]}, [10]} concerning hyperbolic conservation laws and convection-diffusion problems.
| [3] | [
[
478,
481
]
] | https://openalex.org/W1600997507 |
a7d17117-3f2d-4066-801a-28149e845fc0 | Discontinuous Galerkin (dG) finite element methods were introduced in the early 1970s for the numerical solution of first-order hyperbolic problems [1]} and the weak imposition of inhomogeneous boundary conditions for elliptic problems [2]}. In the past several decades, dG methods have enjoyed considerable success as a standard variational framework for the numerical solution of many classes of problems involving partial differential equations (PDEs); see, e.g., monographs [3]}, [4]}, [5]} for reviews of some of the main developments of dG methods.
The interest in dG methods can be attributed to a number of factors, including the great flexibility in dealing with \(hp\) -adaptivity and general shaped elements [6]}, [7]}, [8]},
as well as in solving convection-dominated PDEs; see, e.g., early works [9]}, [10]} concerning hyperbolic conservation laws and convection-diffusion problems.
| [4] | [
[
484,
487
]
] | https://openalex.org/W628726389 |
bfcabf28-8f8a-4268-807d-412c038aa7b6 | Discontinuous Galerkin (dG) finite element methods were introduced in the early 1970s for the numerical solution of first-order hyperbolic problems [1]} and the weak imposition of inhomogeneous boundary conditions for elliptic problems [2]}. In the past several decades, dG methods have enjoyed considerable success as a standard variational framework for the numerical solution of many classes of problems involving partial differential equations (PDEs); see, e.g., monographs [3]}, [4]}, [5]} for reviews of some of the main developments of dG methods.
The interest in dG methods can be attributed to a number of factors, including the great flexibility in dealing with \(hp\) -adaptivity and general shaped elements [6]}, [7]}, [8]},
as well as in solving convection-dominated PDEs; see, e.g., early works [9]}, [10]} concerning hyperbolic conservation laws and convection-diffusion problems.
| [6] | [
[
719,
722
]
] | https://openalex.org/W2116404756 |
d5540343-5dd2-4bb3-a1e6-8fc1c081c4d4 | Discontinuous Galerkin (dG) finite element methods were introduced in the early 1970s for the numerical solution of first-order hyperbolic problems [1]} and the weak imposition of inhomogeneous boundary conditions for elliptic problems [2]}. In the past several decades, dG methods have enjoyed considerable success as a standard variational framework for the numerical solution of many classes of problems involving partial differential equations (PDEs); see, e.g., monographs [3]}, [4]}, [5]} for reviews of some of the main developments of dG methods.
The interest in dG methods can be attributed to a number of factors, including the great flexibility in dealing with \(hp\) -adaptivity and general shaped elements [6]}, [7]}, [8]},
as well as in solving convection-dominated PDEs; see, e.g., early works [9]}, [10]} concerning hyperbolic conservation laws and convection-diffusion problems.
| [8] | [
[
731,
734
]
] | https://openalex.org/W2116404756 |
0080b7dd-e3b1-4448-ab08-dfb6b33f8596 | Discontinuous Galerkin (dG) finite element methods were introduced in the early 1970s for the numerical solution of first-order hyperbolic problems [1]} and the weak imposition of inhomogeneous boundary conditions for elliptic problems [2]}. In the past several decades, dG methods have enjoyed considerable success as a standard variational framework for the numerical solution of many classes of problems involving partial differential equations (PDEs); see, e.g., monographs [3]}, [4]}, [5]} for reviews of some of the main developments of dG methods.
The interest in dG methods can be attributed to a number of factors, including the great flexibility in dealing with \(hp\) -adaptivity and general shaped elements [6]}, [7]}, [8]},
as well as in solving convection-dominated PDEs; see, e.g., early works [9]}, [10]} concerning hyperbolic conservation laws and convection-diffusion problems.
| [10] | [
[
815,
819
]
] | https://openalex.org/W2029017626 |
cf380ab6-7244-478c-8fd6-3aa779f92df3 | Due to missing tools in the analysis, the convergence rate always contains suboptimality in terms of the polynomial degree \(p\) .
In [1]}, the first optimal convergence rate of \(hp\) -dG methods is derived for linear convection problems by using the SUPG stabilisation.
However, the authors provide numerical evidence that the \(hp\) -optimal convergence rate is achieved even without such stabilisation.
In seminal work [2]}, based on (back then) novel optimal approximation results for the \(L^2\) -orthogonal projection, the \(hp\) -optimal convergence rate is derived for dG methods applied to hyperbolic problems, under the technical assumption that the convection field is piecewise linear.
Moreover, whenever the above assumption is violated, the theoretical analysis in [2]} leads to error bounds that are suboptimal in terms of \(p\) by \(3/2\) order.
Such suboptimality is yet not observed in the numerical experiments.
Over the last two decades, the above mentioned technical assumption became standard in \(hp\) -dG methods for convection-diffusion-reaction and hyperbolic problems; see, e.g., [4]}, [5]}, [6]}, [7]}.
It is still an open question, whether the \(p\) -suboptimality for dG methods by \(3/2\) order is true or not in general.
| [6] | [
[
1122,
1125
]
] | https://openalex.org/W2116404756 |
aa26ed01-0eda-416e-9e96-bb3eb4d3248e | Due to missing tools in the analysis, the convergence rate always contains suboptimality in terms of the polynomial degree \(p\) .
In [1]}, the first optimal convergence rate of \(hp\) -dG methods is derived for linear convection problems by using the SUPG stabilisation.
However, the authors provide numerical evidence that the \(hp\) -optimal convergence rate is achieved even without such stabilisation.
In seminal work [2]}, based on (back then) novel optimal approximation results for the \(L^2\) -orthogonal projection, the \(hp\) -optimal convergence rate is derived for dG methods applied to hyperbolic problems, under the technical assumption that the convection field is piecewise linear.
Moreover, whenever the above assumption is violated, the theoretical analysis in [2]} leads to error bounds that are suboptimal in terms of \(p\) by \(3/2\) order.
Such suboptimality is yet not observed in the numerical experiments.
Over the last two decades, the above mentioned technical assumption became standard in \(hp\) -dG methods for convection-diffusion-reaction and hyperbolic problems; see, e.g., [4]}, [5]}, [6]}, [7]}.
It is still an open question, whether the \(p\) -suboptimality for dG methods by \(3/2\) order is true or not in general.
| [7] | [
[
1128,
1131
]
] | https://openalex.org/W2948386797 |
05f7ef6f-1751-418a-a811-2583aca3b99d | The well-posedness of method (REF ) can be found, e.g., in [1]}.
| [1] | [
[
59,
62
]
] | https://openalex.org/W628726389 |
03c27614-2a23-4287-8959-21db6783fe1f | Thanks to assumption (REF ) and an \(\) -polynomial inverse estimate involving bubbles,
see, e.g., [1]}, we deduce
\(\left\Vert \frac{b_j - \mathcal {I}_j b_j}{\sqrt{w_j}} \right\Vert _{\infty ,K} \lesssim K\vert b_j \vert _{W^{2,\infty }(K)}, \quad \quad \Vert \sqrt{w_j} \partial _j \xi \Vert _{0,K} \lesssim \sqrt{p(p+1)} \Vert \xi \Vert _{0,K}.\)
| [1] | [
[
99,
102
]
] | https://openalex.org/W253824085 |
fd39448c-619d-4b61-88cc-28d0f3b91908 | Integrable turbulence is a fascinating topic of nonlinear physics. Integrable turbulence is theoretically and numerically described in the framework of integrable equations amongst which the KdV equation [1]}, [2]}, the Gardner equation [3]}, and the 1D-nonlinear Schrödinger equation [4]}, [5]}. In these systems an infinite number of degrees of freedom can be excited randomly. As such there is no energy transfer between these modes and the word turbulence does not refer to the usual energy cascade between scales. Nevertheless, these systems can exhibit complex random behaviors that require a statistical description. Integrable turbulence applies to many fields of physics: hydrodynamics, optics, and plasmas. [6]}, [7]}, [8]}, [9]}, [10]}, [11]}, [1]}, [13]}, [14]}, [15]}, [16]}.
| [2] | [
[
210,
213
]
] | https://openalex.org/W2963740033 |
6e15c130-6949-48c6-a908-b3493a4ba259 | Integrable turbulence is a fascinating topic of nonlinear physics. Integrable turbulence is theoretically and numerically described in the framework of integrable equations amongst which the KdV equation [1]}, [2]}, the Gardner equation [3]}, and the 1D-nonlinear Schrödinger equation [4]}, [5]}. In these systems an infinite number of degrees of freedom can be excited randomly. As such there is no energy transfer between these modes and the word turbulence does not refer to the usual energy cascade between scales. Nevertheless, these systems can exhibit complex random behaviors that require a statistical description. Integrable turbulence applies to many fields of physics: hydrodynamics, optics, and plasmas. [6]}, [7]}, [8]}, [9]}, [10]}, [11]}, [1]}, [13]}, [14]}, [15]}, [16]}.
| [4] | [
[
285,
288
]
] | https://openalex.org/W2080649718 |
ebe2ea2d-f925-4311-8198-e2b06200ddd5 | Integrable turbulence is a fascinating topic of nonlinear physics. Integrable turbulence is theoretically and numerically described in the framework of integrable equations amongst which the KdV equation [1]}, [2]}, the Gardner equation [3]}, and the 1D-nonlinear Schrödinger equation [4]}, [5]}. In these systems an infinite number of degrees of freedom can be excited randomly. As such there is no energy transfer between these modes and the word turbulence does not refer to the usual energy cascade between scales. Nevertheless, these systems can exhibit complex random behaviors that require a statistical description. Integrable turbulence applies to many fields of physics: hydrodynamics, optics, and plasmas. [6]}, [7]}, [8]}, [9]}, [10]}, [11]}, [1]}, [13]}, [14]}, [15]}, [16]}.
| [7] | [
[
723,
726
]
] | https://openalex.org/W2303595150 |
f83996a6-6079-4d3c-bf4e-4d8efb6186cc | Integrable turbulence is a fascinating topic of nonlinear physics. Integrable turbulence is theoretically and numerically described in the framework of integrable equations amongst which the KdV equation [1]}, [2]}, the Gardner equation [3]}, and the 1D-nonlinear Schrödinger equation [4]}, [5]}. In these systems an infinite number of degrees of freedom can be excited randomly. As such there is no energy transfer between these modes and the word turbulence does not refer to the usual energy cascade between scales. Nevertheless, these systems can exhibit complex random behaviors that require a statistical description. Integrable turbulence applies to many fields of physics: hydrodynamics, optics, and plasmas. [6]}, [7]}, [8]}, [9]}, [10]}, [11]}, [1]}, [13]}, [14]}, [15]}, [16]}.
| [8] | [
[
729,
732
]
] | https://openalex.org/W2340178336 |
0fa1a5d3-2230-4a73-a88a-e24334ece9d5 | Integrable turbulence is a fascinating topic of nonlinear physics. Integrable turbulence is theoretically and numerically described in the framework of integrable equations amongst which the KdV equation [1]}, [2]}, the Gardner equation [3]}, and the 1D-nonlinear Schrödinger equation [4]}, [5]}. In these systems an infinite number of degrees of freedom can be excited randomly. As such there is no energy transfer between these modes and the word turbulence does not refer to the usual energy cascade between scales. Nevertheless, these systems can exhibit complex random behaviors that require a statistical description. Integrable turbulence applies to many fields of physics: hydrodynamics, optics, and plasmas. [6]}, [7]}, [8]}, [9]}, [10]}, [11]}, [1]}, [13]}, [14]}, [15]}, [16]}.
| [13] | [
[
761,
765
]
] | https://openalex.org/W3104785954 |
cf1b01ca-41ff-46ce-a1f7-4bd1d436f0bb | The theory of integrable turbulence in water wave problems is found to be analytically tractable for two “asymptotic” situations.
On the one hand when the waves are of small amplitude the expansion in powers of non linearity yields kinetic equations that model wave resonant interactions. In 2D situations such as for Kadomtsev-Petviashvili type equations [1]} resonant interactions are three-wave resonances. It is known that in the case of the KdV equation the first nontrivial resonances are five-wave interactions but with zero amplitudes [2]}. This result tends to indicate that integrable weak wave turbulence in 1D shallow water cases such as for the KdV equations is precluded.
| [1] | [
[
356,
359
]
] | https://openalex.org/W2080649718 |
d3ed370b-92d3-4ff7-98d8-638f3887a711 | On the other hand when the turbulence can be considered as a collection of solitons with random amplitudes and phases,
kinetic theories of rarefied soliton gas [1]} or dense ones [2]}, [3]}, [4]}, [5]} can be derived . A soliton gas is thus a random state in which solitons behave as quasi-particles due to the fact that their collisions are elastic, only altering relative phases, and thus changing the mean phase speed [1]}. Solitons in the shallow water framework are localized waves which propagate without change of shape due to a balance between linear dispersive effects that tend to flatten out any surface perturbations and nonlinear effects that steepen the fronts. Solitons are at the core of integrable dynamics of the KdV equation.
| [1] | [
[
160,
163
],
[
421,
424
]
] | https://openalex.org/W3146553400 |
34a77ff5-816c-4d43-b8d2-dd086485c972 | On the other hand when the turbulence can be considered as a collection of solitons with random amplitudes and phases,
kinetic theories of rarefied soliton gas [1]} or dense ones [2]}, [3]}, [4]}, [5]} can be derived . A soliton gas is thus a random state in which solitons behave as quasi-particles due to the fact that their collisions are elastic, only altering relative phases, and thus changing the mean phase speed [1]}. Solitons in the shallow water framework are localized waves which propagate without change of shape due to a balance between linear dispersive effects that tend to flatten out any surface perturbations and nonlinear effects that steepen the fronts. Solitons are at the core of integrable dynamics of the KdV equation.
| [2] | [
[
179,
182
]
] | https://openalex.org/W2158432659 |
e1d61888-8d11-47be-8409-dc62d9e1baf5 | On the other hand when the turbulence can be considered as a collection of solitons with random amplitudes and phases,
kinetic theories of rarefied soliton gas [1]} or dense ones [2]}, [3]}, [4]}, [5]} can be derived . A soliton gas is thus a random state in which solitons behave as quasi-particles due to the fact that their collisions are elastic, only altering relative phases, and thus changing the mean phase speed [1]}. Solitons in the shallow water framework are localized waves which propagate without change of shape due to a balance between linear dispersive effects that tend to flatten out any surface perturbations and nonlinear effects that steepen the fronts. Solitons are at the core of integrable dynamics of the KdV equation.
| [3] | [
[
185,
188
]
] | https://openalex.org/W2098476508 |
590da5b5-962b-4f8b-bbf1-b4c8a6ef6d89 | On the other hand when the turbulence can be considered as a collection of solitons with random amplitudes and phases,
kinetic theories of rarefied soliton gas [1]} or dense ones [2]}, [3]}, [4]}, [5]} can be derived . A soliton gas is thus a random state in which solitons behave as quasi-particles due to the fact that their collisions are elastic, only altering relative phases, and thus changing the mean phase speed [1]}. Solitons in the shallow water framework are localized waves which propagate without change of shape due to a balance between linear dispersive effects that tend to flatten out any surface perturbations and nonlinear effects that steepen the fronts. Solitons are at the core of integrable dynamics of the KdV equation.
| [4] | [
[
191,
194
]
] | https://openalex.org/W3188857046 |
7aaaddca-11a2-4423-81c8-5308fd38f937 | On the other hand when the turbulence can be considered as a collection of solitons with random amplitudes and phases,
kinetic theories of rarefied soliton gas [1]} or dense ones [2]}, [3]}, [4]}, [5]} can be derived . A soliton gas is thus a random state in which solitons behave as quasi-particles due to the fact that their collisions are elastic, only altering relative phases, and thus changing the mean phase speed [1]}. Solitons in the shallow water framework are localized waves which propagate without change of shape due to a balance between linear dispersive effects that tend to flatten out any surface perturbations and nonlinear effects that steepen the fronts. Solitons are at the core of integrable dynamics of the KdV equation.
| [5] | [
[
197,
200
]
] | https://openalex.org/W3150535889 |
e452cbef-b1c3-4e66-847e-d001d918ae23 | Empirical confirmation that soliton gases can be generated were given in optics [1]}, for deep water waves [2]} and for shallow water wave motion [3]}, [4]}. In the experiments energy dissipation cannot be avoided and this seems at first glance strongly incompatible with the concept of integrability. However, [3]} observed that due to a large scale separation between the nonlinear scale related to the (short) duration of soliton collisions and the (long) dissipative time scale, the dynamics is overall ruled by integrability. A stationary random soliton gas in a long wave flume in shallow water conditions can thus be sustained with continuous energy input by the wavemaker. Even though not labeled as soliton gases some 1D flume experiments of [6]}, in which the wavemaker has a sinusoidal displacement, lead to random wave motions. Therefore, the role of the wavemaker in the outcome of these random stationnary wave states needs to be explored.
| [2] | [
[
107,
110
]
] | https://openalex.org/W3039663288 |
61857c2d-a49c-4f6e-9e3b-6b79f025e50b | Empirical confirmation that soliton gases can be generated were given in optics [1]}, for deep water waves [2]} and for shallow water wave motion [3]}, [4]}. In the experiments energy dissipation cannot be avoided and this seems at first glance strongly incompatible with the concept of integrability. However, [3]} observed that due to a large scale separation between the nonlinear scale related to the (short) duration of soliton collisions and the (long) dissipative time scale, the dynamics is overall ruled by integrability. A stationary random soliton gas in a long wave flume in shallow water conditions can thus be sustained with continuous energy input by the wavemaker. Even though not labeled as soliton gases some 1D flume experiments of [6]}, in which the wavemaker has a sinusoidal displacement, lead to random wave motions. Therefore, the role of the wavemaker in the outcome of these random stationnary wave states needs to be explored.
| [3] | [
[
146,
149
],
[
311,
314
]
] | https://openalex.org/W2942459581 |
4054b07c-1333-4e6a-90a3-dffab0b0547c | \(50\,\) years have elapsed between the first theoretical description of soliton gases by [1]} and the first hydrodynamic experiments. A possible reason is the requirement of highly resolved instruments to capture the space-time evolution of a random state. Another more fundamental issue relates to the difference between an infinite or periodic domain setting, usually used in theoretical approaches, and finite length experimental set-ups. This difference also combines with how initial conditions are easily prescribed in theory and numerics while boundary conditions are most of the time the only options at least in hydrodynamic experiments. A recent notable exception are the deep water soliton gas experiments by Suret et al. [2]} in which the Inverse Scattering Transform for the 1D Schrödinger equation is used to compute boundary conditions in a very long flume. In these experiments an ensemble of random spectral values associated to solitons are prescribed which then evolve towards a soliton gas.
| [1] | [
[
91,
94
]
] | https://openalex.org/W3146553400 |
5665867a-b6cf-42e2-9de7-baf8b55d2bb9 | \(50\,\) years have elapsed between the first theoretical description of soliton gases by [1]} and the first hydrodynamic experiments. A possible reason is the requirement of highly resolved instruments to capture the space-time evolution of a random state. Another more fundamental issue relates to the difference between an infinite or periodic domain setting, usually used in theoretical approaches, and finite length experimental set-ups. This difference also combines with how initial conditions are easily prescribed in theory and numerics while boundary conditions are most of the time the only options at least in hydrodynamic experiments. A recent notable exception are the deep water soliton gas experiments by Suret et al. [2]} in which the Inverse Scattering Transform for the 1D Schrödinger equation is used to compute boundary conditions in a very long flume. In these experiments an ensemble of random spectral values associated to solitons are prescribed which then evolve towards a soliton gas.
| [2] | [
[
735,
738
]
] | https://openalex.org/W3039663288 |
fd5dcfd8-e130-4e30-a87f-20dca440ad79 | The constraints on the experiments mentioned above, require to find other routes to the generation of integrable turbulence and soliton gases, an aspect investigated in the present work. In this context the question of the statistical properties of the stationary state of integrable turbulence needs to be addressed since it remains largely open and was mostly investigated by numerical studies [1]}, [2]}, [3]}, [4]}.
| [2] | [
[
402,
405
]
] | https://openalex.org/W2963740033 |
03774173-49de-4477-8de2-474000f218cc | The details of the experimental setup and data analysis tools can be found in [1]} and some aspects are discussed in [2]}, [3]} as well. We only recall here the main features of the experimental set-up.
| [2] | [
[
117,
120
]
] | https://openalex.org/W2942459581 |
08e0f71a-9c7e-48b6-a79e-872701bf55e0 | Our goal is to obtain integrable turbulence steady states such as soliton gases. Thus, one would like to generate as many solitons as possible with the wavemaker. In order to ultimately sustain a large number of solitons in the flume we take advantage i) of the well-known fission phenomenon in shallow water of a sine wave train into solitons as evidenced numerically by Zabusky & Kruskal [1]} and observed experimentally by Zabusky & Galvin [2]} and in a more comprehensive way by Trillo et. al [3]} and ii) of the amplitude amplification of the non-linear modes by their interactions with the moving wavemaker.
Mention should be made of the experiments by [4]} who, by slightly detuning the wavemaker motion with respect to the periodic longitudinal seiching mode of the channel, were able to find a route towards the generation of integrable turbulence.
| [1] | [
[
390,
393
]
] | https://openalex.org/W2050639649 |
2cfbe990-b2a0-4149-b846-4bf32fa9030d |
ridge a: signature of weak dispersive shallow water waves following the Airy dispersion relation
\(\omega ^2 = g \, k \, \tanh (k \, h).\)
These dispersive waves originate from the bound waves of the wavemaker monochromatic sinusoidal wave forcing and from the weak non integrable effects during soliton collisions [1]}, [2]};
ridge b: signature of shallow water non-linear waves which Fourier modes all travel at a velocity close to \(c_0 = \sqrt{g \, h}\) .
ridges c: signature of transverse waves. The uni-nodal transverse wave would be at a frequency of \(0.9\,\) Hz.
The energy of the uni-nodal transverse waves is three orders of magnitude smaller than that of longitudinal waves.
| [2] | [
[
323,
326
]
] | https://openalex.org/W2942459581 |
bcc85d17-b805-4693-a0c2-c982066fd1c2 | The case described in Fig. REF and REF and analyzed in [1]} is typically an example of integrable turbulence containing a significant number of solitons. Solitons are responsible for “ridge b” of Fig. REF and they also clearly leave straight line signatures in the \((x,t)\) plane of Fig. REF . This was discussed in detail in [2]}.
| [1] | [
[
57,
60
]
] | https://openalex.org/W2942459581 |
2030a331-09fc-4d62-8cde-a9bbe8ee5dc1 | Gardner et al. [1]} made a significant leap forward with the Direct Scattering Transform (DST) of the KdV equation that extracts the spectrum of the associated Schrödinger equation which potential is the nonlinear wave signal to be analyzed. This method decomposes a time series into nonlinear solitonic modes that evolve independently in time and space without change of shape along with radiating shallow water weakly nonlinear waves. Once the time independent spectrum of nonlinear modes is determined, the signal can be reconstructed at any time by the Inverse Scattering Transform (IST).
DST theory for the case of a localized initial condition [1]} differs from the much more complex case of the spatially periodic initial condition known as the finite band theory [3]}, [4]}, [5]}, [6]}.
In our experimental case we are more concerned with the periodic case due to our configuration in which the waves propagate back and forth in the flume. Although the waves in the present experiments are not in most cases periodic, the wave motion is however confined in space and does not decay at infinity. This approach has been used by [7]} to analyze oceanic field data.
| [1] | [
[
15,
18
],
[
650,
653
]
] | https://openalex.org/W2030669316 |
106232d0-808c-408d-bee3-46dc11e7b8a8 | Gardner et al. [1]} made a significant leap forward with the Direct Scattering Transform (DST) of the KdV equation that extracts the spectrum of the associated Schrödinger equation which potential is the nonlinear wave signal to be analyzed. This method decomposes a time series into nonlinear solitonic modes that evolve independently in time and space without change of shape along with radiating shallow water weakly nonlinear waves. Once the time independent spectrum of nonlinear modes is determined, the signal can be reconstructed at any time by the Inverse Scattering Transform (IST).
DST theory for the case of a localized initial condition [1]} differs from the much more complex case of the spatially periodic initial condition known as the finite band theory [3]}, [4]}, [5]}, [6]}.
In our experimental case we are more concerned with the periodic case due to our configuration in which the waves propagate back and forth in the flume. Although the waves in the present experiments are not in most cases periodic, the wave motion is however confined in space and does not decay at infinity. This approach has been used by [7]} to analyze oceanic field data.
| [3] | [
[
771,
774
]
] | https://openalex.org/W1973107894 |
34421b73-5ceb-4610-9098-1baa7cfae64e | Gardner et al. [1]} made a significant leap forward with the Direct Scattering Transform (DST) of the KdV equation that extracts the spectrum of the associated Schrödinger equation which potential is the nonlinear wave signal to be analyzed. This method decomposes a time series into nonlinear solitonic modes that evolve independently in time and space without change of shape along with radiating shallow water weakly nonlinear waves. Once the time independent spectrum of nonlinear modes is determined, the signal can be reconstructed at any time by the Inverse Scattering Transform (IST).
DST theory for the case of a localized initial condition [1]} differs from the much more complex case of the spatially periodic initial condition known as the finite band theory [3]}, [4]}, [5]}, [6]}.
In our experimental case we are more concerned with the periodic case due to our configuration in which the waves propagate back and forth in the flume. Although the waves in the present experiments are not in most cases periodic, the wave motion is however confined in space and does not decay at infinity. This approach has been used by [7]} to analyze oceanic field data.
| [4] | [
[
777,
780
]
] | https://openalex.org/W2118575947 |
8f74a0fb-a471-4bdf-9804-1dbfca1d0935 | Gardner et al. [1]} made a significant leap forward with the Direct Scattering Transform (DST) of the KdV equation that extracts the spectrum of the associated Schrödinger equation which potential is the nonlinear wave signal to be analyzed. This method decomposes a time series into nonlinear solitonic modes that evolve independently in time and space without change of shape along with radiating shallow water weakly nonlinear waves. Once the time independent spectrum of nonlinear modes is determined, the signal can be reconstructed at any time by the Inverse Scattering Transform (IST).
DST theory for the case of a localized initial condition [1]} differs from the much more complex case of the spatially periodic initial condition known as the finite band theory [3]}, [4]}, [5]}, [6]}.
In our experimental case we are more concerned with the periodic case due to our configuration in which the waves propagate back and forth in the flume. Although the waves in the present experiments are not in most cases periodic, the wave motion is however confined in space and does not decay at infinity. This approach has been used by [7]} to analyze oceanic field data.
| [5] | [
[
783,
786
]
] | https://openalex.org/W1661143721 |
d1990f89-ad4c-490c-af55-d7d6f20f4336 | Gardner et al. [1]} made a significant leap forward with the Direct Scattering Transform (DST) of the KdV equation that extracts the spectrum of the associated Schrödinger equation which potential is the nonlinear wave signal to be analyzed. This method decomposes a time series into nonlinear solitonic modes that evolve independently in time and space without change of shape along with radiating shallow water weakly nonlinear waves. Once the time independent spectrum of nonlinear modes is determined, the signal can be reconstructed at any time by the Inverse Scattering Transform (IST).
DST theory for the case of a localized initial condition [1]} differs from the much more complex case of the spatially periodic initial condition known as the finite band theory [3]}, [4]}, [5]}, [6]}.
In our experimental case we are more concerned with the periodic case due to our configuration in which the waves propagate back and forth in the flume. Although the waves in the present experiments are not in most cases periodic, the wave motion is however confined in space and does not decay at infinity. This approach has been used by [7]} to analyze oceanic field data.
| [7] | [
[
1134,
1137
]
] | https://openalex.org/W1984085074 |
e204fb89-a3c4-470a-afce-57fe8f5cdc11 | We implemented the Periodic Scattering Transform algorithm (PST) for the KdV equation developed by Osborne [1]}, [2]}, [3]} in a program to compute the nonlinear spectrum of the experimental free surface records. In the periodic case, the non linear modes are described by hyper-elliptic functions. These waves similar to cnoidal waves are characterized by the so-called spectral modulus \(m\) which quantifies the level of nonlinearity and which is an output of the PST [2]}. For a vanishing small modulus the modes are close to sine waves. For \(m\) very close to 1, the modes appear as localized pulses in the periodic box that resemble solitons. The delineation between solitonic modes and radiation modes thus breaks down to the choice of the threshold modulus. This threshold modulus also defines the reference depth \(h_{ref}\) [1]}, [2]}, [3]} on which these solitonic modes propagate. Osborne [1]} suggests that modes with \(m>0.99\) can be considered as solitons a definition also used by [9]}. We will not discuss here the details of our implementation of the PST which is described in [10]} with different validation cases and an analysis of various limitations since this method has been validated various times [11]}, [9]}. The main output of the PST is the spectrum that lists the nonlinear modes and their moduli.
| [1] | [
[
107,
110
],
[
838,
841
],
[
905,
908
]
] | https://openalex.org/W366502457 |
cff6340d-026f-4599-a0b8-7b601a118cd4 | One of the relevant non-dimensional number in the shallow water context is the Ursell number [1]}.
It is even the only dimensionless number for KdV unidirectional wave motion dynamics.
The Ursell number is the ratio in order of magnitude of the nonlinear to dispersive terms of the KdV equation. The dimensional version of the KdV equation is,
\(\partial _t \eta + c_0 \, \partial _x \eta + \frac{3 \, c_0}{2 \, h} \, \eta \, \partial _x \eta + \frac{h^2 \, c_0}{6} \, \partial _{xxx} \eta = 0. \)
| [1] | [
[
93,
96
]
] | https://openalex.org/W2030188624 |
ad6415bc-34cb-44b5-bb59-58cb824dc748 | Equation (REF ) shows that for very small \(U\) the equation becomes linear and dispersive, the so-called Airy equation.
Under this condition the forcing wave remains linear but disperses (different frequency components propagate at different speeds) as it propagates and no soliton emerges. For large values of \(U\) equation (REF ) becomes non-linear of the Burgers type. A sinusoidal forcing wave that fulfills such condition will undergo nonlinear steepening up to the gradient catastrophe producing a steep front face (shock wave). At that point the wave front face characteristic length is small and dispersion comes into play. Dispersion forces the shock wave to fission into a train of solitons [1]}, [2]}, [3]}. The Ursell number of the forcing wave therefore indicates how many solitons are expected [3]} . Since our set-up does not allow for initial condition recurrence to take place because of the end wall reflection, the Ursell number also measures how disorganized the regime is. Indeed, solitons will reflect back on the end wall and the wavemaker, interacting with others, generating phase shifts and therefore possibly disorganizing the initial periodicity.
<FIGURE> | [1] | [
[
705,
708
]
] | https://openalex.org/W2050639649 |
d438e96c-172f-4aad-a87a-3ba40cfc26bf | However, we observed that the sole value of the forcing wave Ursell number \(U\) is insufficient to discriminate between the periodic and random states. Fig. REF shows the frequency spectra of 9 experiments for the left-running component of the wave motion at \(x=40\,\) m which is the best compromise between soliton separation after fission and dissipation. All these experiments have very close Ursell number values of \(U = 0.53\) , but distinct values of the frequency \(f\) and amplitude \(a\) . Signatures of different stationary states are observed from very periodic ones at low forcing frequency and low forcing amplitude to random ones at high forcing frequency and high forcing amplitude and going through continuously varying power spectra shape in between. This observation shows that another dimensionless parameter must be taken into account to sort the different states out. Fig. REF shows the corresponding distribution of non-linear mode amplitudes given by the PST. The periodic cases (3 top subplots of Fig. REF and correspondingly in Fig. REF ) exhibit sets of non-linear modes of nearly constant amplitudes indicating that these cases remain strongly organized. For instance the case \(f=0.18\,\) Hz for a 60 s wave motion recording corresponds to roughly 10 periods of the forcing wave for which the PST gives 10 soliton modes of \(0.15\,\) cm amplitudes and hardly any other modes. A small soliton of \(0.15\,\) cm amplitude is locked to each wave period. By contrast the soliton modes in the random cases are much more numerous and their amplitudes distributed over a large range indicating indeed that these cases are random. These states (3 bottom subplots of Fig. REF and correspondingly in Figure REF ) are considered to be what is called integrable turbulence [1]}. Discussion of the amplitude distribution is post-poned to section .
<FIGURE> | [1] | [
[
1797,
1800
]
] | https://openalex.org/W2942459581 |
086fee98-0e9b-45cb-95ad-29acd8071272 | We conclude from these observations that another way to synthetically discriminate the various regimes, within our experimental framework, is to compute the frequency spectrum as given in Fig. REF . In the periodic case, the spectrum is mostly made of narrow peaks with the fundamental peak at the forcing frequency and the other peaks corresponding to higher harmonics. In the random case the spectrum is quite distinct and it is seen to be continuous
with a flat plateau [1]}, [2]} that extends from the forcing frequency down to the smallest resolved frequency.
At frequencies higher than the forcing frequency the spectrum decays exponentially.
At this point, a question arises on how to quantify the state of the wave field. At this end, we define a dimensionless randomness index \(I_m\) based on the shape of the power spectrum:
\(I_m=\log _{10}\left(\frac{E_{forcing}}{E_{av}}\right) \)
| [2] | [
[
479,
482
]
] | https://openalex.org/W2942459581 |
ed66017e-8e15-4a5a-9ecf-dce2ac0b5326 | In this section statistical distributions and moments of the experimental integrable turbulence are presented. Statistical information complements PST analysis. PST in our present study is a key tool to assess the existence of a soliton gas and characterize the modal content of the gas. Nonetheless integrable turbulence encompasses also shallow water situations where solitonic modes coexist with radiation modes.
As discussed in the introduction the literature on the statistical description of integrable turbulence is sparse while
such situation can be present in field measurements [1]}. A noticeable exception is the numerical study by [2]} on KdV random wave fields, a reference study on these aspects.
| [1] | [
[
588,
591
]
] | https://openalex.org/W1984085074 |
1c887344-279b-4c32-b11e-98c8f4790da1 | In a periodic box and with an initial condition made of a sine wave, Zabuski & Kruskal [1]} numerically predict a recurrence, i.e. the fact that the wave system retrieves in a finite time a state very close to the initial condition [1]}. In our experimental setup, it is not possible to start from an arbitrary initial condition which is not rest. What can only be achieved in a controlled way, is to start wave forcing at one end with a wavemaker and a body of water at rest in the entire flume.
| [1] | [
[
87,
90
],
[
232,
235
]
] | https://openalex.org/W2050639649 |
19758ae5-9c66-4b83-a41d-ea1755f0f377 | Our experimental set-up differs from the integrable framework of Zabusky & Kruskal [1]} on various points. The most obvious is dissipation that imposes some continuous energy flux input for the wave motion to possibly reach a statistically stationary wave regime. However, the time scales of dissipation, well represented by the Keulegan law, are much larger than those involving soliton interactions which suggest that integrability still holds locally. Indeed, we show that once a soliton gas is formed dissipation slowly alters the amplitude distribution but not to the point where it would be obliterated in a flume length propagation time.
| [1] | [
[
83,
86
]
] | https://openalex.org/W2050639649 |
aca6b0a2-0ae3-48c4-8e94-2f3fcbe40f60 | where \(\nu _0=310\) MHz. This is substantially larger than what can be explained by the observed counts of radio-emitting galaxies [1]}. This excess is consistent across several independent measurements, and its origin remains unknown [2]}, [3]}, [4]}. In what follows we assume that this is indeed a cosmic radio background that is of extragalactic origin, consistent with the non-detection of bright radio halos around galaxies similar to ours [5]}. Its existence at high redshift would have important implications for searches for redshifted 21cm emission from the epoch of the first stars [6]}.
Assuming that this radio background is truly cosmological and isotropic, this same background should be present in galaxy clusters.
| [4] | [
[
249,
252
]
] | https://openalex.org/W3111523546 |
0f5270f0-9a3a-4d82-a7ee-2897bf77509c | The predicted radio SZ has a steep scaling towards low frequencies (see Fig. REF ). As such, it could become an important foreground for standard 21cm fluctuation measurements. The signal is expected to spatially correlate with the standard Compton-\(y\) map and therefore can be modelled by combining SZ and 21cm measurements. This is of particular interest to radio experiments such as LOFAR [1]} and the Square Kilometer Array [2]}, and CMB experiments like the Simons Observatory [3]} and CMB-S4 [4]}, which all promise high resolution wide-area maps of the sky.
| [3] | [
[
485,
488
]
] | https://openalex.org/W2888259247 |
e1bebd61-f036-43cd-89ba-e5430d278e1e | The predicted radio SZ has a steep scaling towards low frequencies (see Fig. REF ). As such, it could become an important foreground for standard 21cm fluctuation measurements. The signal is expected to spatially correlate with the standard Compton-\(y\) map and therefore can be modelled by combining SZ and 21cm measurements. This is of particular interest to radio experiments such as LOFAR [1]} and the Square Kilometer Array [2]}, and CMB experiments like the Simons Observatory [3]} and CMB-S4 [4]}, which all promise high resolution wide-area maps of the sky.
| [4] | [
[
501,
504
]
] | https://openalex.org/W2958360979 |
bc06101c-2bd6-49c5-9623-3d88d01e479c | Researchers have investigated the hierarchical RL problem under various settings. Existing theoretical analyses [1]}, [2]}, [3]}, [4]} typically assume that the options are given. As a result, only the high-level policy needs to be learned. Recent advances in deep hierarchical RL (e.g., [5]}) focus on concurrently learning the full options framework, but still the initialization of the options is critical. A promising practical approach is to learn an initial hierarchical policy from expert demonstrations. Then, deep hierarchical RL algorithms can be applied for policy improvement. The former step is named as Hierarchical Imitation Learning (HIL).
| [1] | [
[
112,
115
]
] | https://openalex.org/W2133458291 |
568e6b45-a590-4ebe-a9e0-5d8f49a852fd | Researchers have investigated the hierarchical RL problem under various settings. Existing theoretical analyses [1]}, [2]}, [3]}, [4]} typically assume that the options are given. As a result, only the high-level policy needs to be learned. Recent advances in deep hierarchical RL (e.g., [5]}) focus on concurrently learning the full options framework, but still the initialization of the options is critical. A promising practical approach is to learn an initial hierarchical policy from expert demonstrations. Then, deep hierarchical RL algorithms can be applied for policy improvement. The former step is named as Hierarchical Imitation Learning (HIL).
| [5] | [
[
288,
291
]
] | https://openalex.org/W2964227312 |
12576e89-c067-4f1d-8010-74e940003f3b | In this paper, we investigate HIL from a theoretical perspective. Our problem formulation is concise while retaining the essential difficulty of HIL: we need to learn a complete hierarchical policy from an unsegmented sequence of state-action pairs. Under this setting, HIL becomes an inference problem in a latent variable model. Such a transformation was first proposed by [1]}, where the Expectation-Maximization (EM) algorithm [2]} was applied for policy learning. Empirical results for this algorithm and its gradient variants [3]}, [4]} demonstrate good performance, but the theoretical analysis remains open. By bridging this gap, we aim to solidify the foundation of HIL and provide some high level guidance for its practice.
| [1] | [
[
375,
378
]
] | https://openalex.org/W2498991332 |
acea0ad9-2e8f-42cf-8179-92601bcdd74a | In this paper, we investigate HIL from a theoretical perspective. Our problem formulation is concise while retaining the essential difficulty of HIL: we need to learn a complete hierarchical policy from an unsegmented sequence of state-action pairs. Under this setting, HIL becomes an inference problem in a latent variable model. Such a transformation was first proposed by [1]}, where the Expectation-Maximization (EM) algorithm [2]} was applied for policy learning. Empirical results for this algorithm and its gradient variants [3]}, [4]} demonstrate good performance, but the theoretical analysis remains open. By bridging this gap, we aim to solidify the foundation of HIL and provide some high level guidance for its practice.
| [3] | [
[
532,
535
]
] | https://openalex.org/W2605016475 |
ad2951cc-3942-41c1-839c-ab2e96bfe99f | Due to its intrinsic difficulty, existing works on HIL typically consider its easier variants for practicality. If the expert options are observed, standard imitation learning algorithms can be applied to learn the high and low level policies separately [1]}. If those are not available, a popular idea [2]}, [3]}, [4]}, [5]} is to first divide the expert demonstration into segments using domain knowledge or heuristics, learn the individual option corresponding to each segment, and finally learn the high level policy. With additional supervision, these steps can be unified [6]}. In this regard, the EM approach [7]}, [8]}, [9]} is this particular idea pushed to an extreme: without any other forms of supervision, we simultaneously segment the demonstration and learn from it, by exploiting the latent variable structure.
| [7] | [
[
616,
619
]
] | https://openalex.org/W2498991332 |
ce670411-341d-4cff-b992-657e6efbe938 | Due to its intrinsic difficulty, existing works on HIL typically consider its easier variants for practicality. If the expert options are observed, standard imitation learning algorithms can be applied to learn the high and low level policies separately [1]}. If those are not available, a popular idea [2]}, [3]}, [4]}, [5]} is to first divide the expert demonstration into segments using domain knowledge or heuristics, learn the individual option corresponding to each segment, and finally learn the high level policy. With additional supervision, these steps can be unified [6]}. In this regard, the EM approach [7]}, [8]}, [9]} is this particular idea pushed to an extreme: without any other forms of supervision, we simultaneously segment the demonstration and learn from it, by exploiting the latent variable structure.
| [8] | [
[
622,
625
]
] | https://openalex.org/W2605016475 |
b67af927-f3dc-4673-bd19-de48b385aaac | Recent ideas on EM algorithms [1]}, [2]}, [3]}, [4]} focus on the convergence to the true parameter directly, relying on an instrumental object named as the population EM algorithm. It has the same two-stage iterative procedure as the standard EM algorithm, but its \(Q\) -function, the maximization objective in the M-step, is defined as the infinite sample limit of the finite sample \(Q\) -function. Under regularity conditions, the population EM algorithm converges to the true parameter. The standard EM algorithm is then analyzed as its perturbed version, converging with high probability to a norm ball around the true parameter. The main advantage of this approach is that the true parameter usually has a large basin of attraction in the population EM algorithm. Therefore, the requirement on initialization is less stringent. See [3]} for an illustration.
| [1] | [
[
30,
33
]
] | https://openalex.org/W2962737134 |
41fccfa2-0804-4bb4-b4f6-81c61f988db3 | The \(Q\) -function adopted in the population EM algorithm is named as the population \(Q\) -function. To properly define such a quantity, the stochastic convergence of the finite sample \(Q\) -function needs to be constructed. When the samples are i.i.d., such as in Gaussian Mixture Models (GMMs) [1]}, [2]}, [3]}, the required convergence follows directly from the law of large numbers. However, this argument is less straightforward in time-series models such as Hidden Markov Models (HMMs) and the model considered in HIL. For HMMs, [4]} showed that the expectation of the \(Q\) -function converges, but both the stochastic convergence analysis and the analytical expression of the population \(Q\) -function are not provided. The missing techniques could be borrowed from a body of work [5]}, [6]}, [7]}, [8]} analyzing the asymptotic behavior of HMMs. Most notably, [7]} provided a rigorous treatment of the population EM algorithm via sufficient statistics, assuming the HMM is parameterized by an exponential family.
| [1] | [
[
299,
302
]
] | https://openalex.org/W2962737134 |
95943076-33c3-4a11-8c72-58f096324f9d | The \(Q\) -function adopted in the population EM algorithm is named as the population \(Q\) -function. To properly define such a quantity, the stochastic convergence of the finite sample \(Q\) -function needs to be constructed. When the samples are i.i.d., such as in Gaussian Mixture Models (GMMs) [1]}, [2]}, [3]}, the required convergence follows directly from the law of large numbers. However, this argument is less straightforward in time-series models such as Hidden Markov Models (HMMs) and the model considered in HIL. For HMMs, [4]} showed that the expectation of the \(Q\) -function converges, but both the stochastic convergence analysis and the analytical expression of the population \(Q\) -function are not provided. The missing techniques could be borrowed from a body of work [5]}, [6]}, [7]}, [8]} analyzing the asymptotic behavior of HMMs. Most notably, [7]} provided a rigorous treatment of the population EM algorithm via sufficient statistics, assuming the HMM is parameterized by an exponential family.
| [2] | [
[
305,
308
]
] | https://openalex.org/W2963626260 |
d0cd4757-56a8-426c-a733-1e0db2d7e288 | The \(Q\) -function adopted in the population EM algorithm is named as the population \(Q\) -function. To properly define such a quantity, the stochastic convergence of the finite sample \(Q\) -function needs to be constructed. When the samples are i.i.d., such as in Gaussian Mixture Models (GMMs) [1]}, [2]}, [3]}, the required convergence follows directly from the law of large numbers. However, this argument is less straightforward in time-series models such as Hidden Markov Models (HMMs) and the model considered in HIL. For HMMs, [4]} showed that the expectation of the \(Q\) -function converges, but both the stochastic convergence analysis and the analytical expression of the population \(Q\) -function are not provided. The missing techniques could be borrowed from a body of work [5]}, [6]}, [7]}, [8]} analyzing the asymptotic behavior of HMMs. Most notably, [7]} provided a rigorous treatment of the population EM algorithm via sufficient statistics, assuming the HMM is parameterized by an exponential family.
| [5] | [
[
793,
796
]
] | https://openalex.org/W4250389103 |
c2d7905c-2371-4ee1-8cd6-1ebf8f9ca1ee | The \(Q\) -function adopted in the population EM algorithm is named as the population \(Q\) -function. To properly define such a quantity, the stochastic convergence of the finite sample \(Q\) -function needs to be constructed. When the samples are i.i.d., such as in Gaussian Mixture Models (GMMs) [1]}, [2]}, [3]}, the required convergence follows directly from the law of large numbers. However, this argument is less straightforward in time-series models such as Hidden Markov Models (HMMs) and the model considered in HIL. For HMMs, [4]} showed that the expectation of the \(Q\) -function converges, but both the stochastic convergence analysis and the analytical expression of the population \(Q\) -function are not provided. The missing techniques could be borrowed from a body of work [5]}, [6]}, [7]}, [8]} analyzing the asymptotic behavior of HMMs. Most notably, [7]} provided a rigorous treatment of the population EM algorithm via sufficient statistics, assuming the HMM is parameterized by an exponential family.
| [6] | [
[
799,
802
]
] | https://openalex.org/W2964126880 |
a0d1acd7-4f7f-477c-9a61-d395d5f5d69f | Finally, apart from the EM algorithm, a separate line of research [1]}, [2]} applies spectral methods for tractable inference in latent variable models. However, such methods are mainly complementary to the EM algorithm since better performance can usually be obtained by initializing the EM algorithm with the solution of the spectral methods [3]}.
| [1] | [
[
66,
69
]
] | https://openalex.org/W2105724942 |
bff40df6-c965-498a-bec7-76d16b9a3a90 | In this paper, we establish the first known performance guarantee for a HIL algorithm that only observes primitive state-action pairs. Specifically, we first fix and reformulate the original EM approach by [1]} in a rigorous manner. The lack of mixing is identified as a technical difficulty in learning the standard options framework, and a novel options with failure framework is proposed to circumvent this issue.
| [1] | [
[
206,
209
]
] | https://openalex.org/W2498991332 |
766c3f1b-35a0-4a09-9f3d-bccebc340fa7 | Inspired by [1]} and [2]}, the population version of our algorithm is analyzed as an intermediate step. We prove that if the expert policy can be parameterized by the options with failure framework, then, under regularity conditions, the population version algorithm converges to the true parameter, and the finite sample version converges with high probability to a norm ball around the true parameter. Our analysis directly constructs the stochastic convergence of the finite sample \(Q\) -function, and an analytical expression of the resulting population \(Q\) -function is provided. Finally, we qualitatively validate our theoretical results using a numerical example.
| [1] | [
[
12,
15
]
] | https://openalex.org/W2962737134 |
d28e065f-d8f2-4090-a425-0aa04e1d51b5 | The options with failure framework is adopted to simplify the construction of the mixing condition (Lemma REF ). It is possible that our analysis could be extended to learn the standard options framework. In that case, instead of constructing the usual one step mixing condition, one could target the multi-step mixing condition similar to [1]}.
| [1] | [
[
340,
343
]
] | https://openalex.org/W4250389103 |
c6efe7a9-b412-49de-8928-732778119dd0 | Two comments need to made here. First, it is common in practice to observe not one, but a set of independent observation sequences. In that case, the problem essentially becomes easier. Second, the cardinality of the option space and the parameterization of the expert policy are usually unknown. A popular solution is to assume an expressive parameterization (e.g., a neural network) in the algorithm and select \(\textrm {card}()\) through cross-validation. Theoretical analysis of EM under this setting is challenging, even when samples are i.i.d. [1]}, [2]}. Therefore, we only consider the domain of correct-specification.
| [1] | [
[
552,
555
]
] | https://openalex.org/W3113313164 |
7268d357-99f7-4cc9-a0bb-82a49d832592 | Adopting the EM approach, we present Algorithm for the estimation of \(\theta ^*\) . It reformulates the algorithm by [1]} in a rigorous manner, and an error in the latter is fixed: when defining the posterior distribution of latent variables, at any time \(t<T\) , the original algorithm neglects the dependency of future states \(S_{t+1:T}\) on the current option \(O_t\) . A detailed discussion is provided in Appendix REF .
| [1] | [
[
119,
122
]
] | https://openalex.org/W2498991332 |
31d90fbb-99e6-4f06-a02a-f20bc8fcd81b | Although we require finite state and action space for our theoretical analysis, Algorithm can be readily generalized to continuous \(\) and \(\) : we only need to replace \(\pi _{lo}\) by a density function. However, generalization to continuous option space requires a substantially different algorithm. The forward-backward smoothing procedure in Theorem REF involves integrals rather than sums, and Sequential Monte Carlo (SMC) techniques need to be applied. Fortunately, it is widely accepted that a finite option space is reasonable in the options framework, since the options need to be distinct and separate [1]}.
| [1] | [
[
619,
622
]
] | https://openalex.org/W2111967991 |
c41f11b1-2e34-4c9f-9053-a765c7713194 | Our analysis of Algorithm has the following structure. We first prove the stochastic convergence of the \(Q\) -function \(Q_{\mu ,T}(\theta ^{\prime }|\theta )\) to a population \(Q\) -function \(\bar{Q}(\theta ^{\prime }|\theta )\) , leading to a well-posed definition of the population version algorithm. This step is our major theoretical contribution. With additional assumptions, the first-order stability condition is constructed, and techniques in [1]} can be applied to show the convergence of the population version algorithm. The remaining step is to analyze Algorithm as a perturbed form of its population version, which requires a high probability bound on the distance between their parameter updates. We can establish the strong consistency of the parameter update of Algorithm as an estimator of the parameter update of the population version algorithm. Therefore, the existence of such a high probability bound can be proved for large enough \(T\) . However, the analytical expression of this bound requires knowledge of the specific parameterization of \(\lbrace \bar{\pi }_{hi},\pi _{lo}, \pi _b\rbrace \) , which is not available in this general context of discussion.
| [1] | [
[
457,
460
]
] | https://openalex.org/W2962737134 |
d0bce762-5b2a-4745-aeac-de75d7eb2cf0 | [Proof Sketch]
The main difficulty of the proof is that, \(Q_{\mu ,T}(\theta ^{\prime }|\theta )\) defined in (REF ) is (roughly) the average of \(T\) terms, with each term dependent on the entire observation sequence; as \(T\rightarrow \infty \) , all the terms keep changing such that the law of large numbers cannot be applied directly. As a solution, we approximate \(\gamma ^{\theta }_{\mu ,t|T}\) and \(\tilde{\gamma }^{\theta }_{\mu ,t|T}\) with smoothing distributions in an infinitely extended graphical model independent of \(T\) , resulting in an approximated \(Q\) -function (still depends on \(T\) ). The techniques adopted in this step are analogous to Markovian decomposition and uniform forgetting in the HMM literature [1]}, [2]}. The limiting behavior of the approximated \(Q\) -function is the same as that of \(Q_{\mu ,T}(\theta ^{\prime }|\theta )\) , since their difference vanishes as \(T\rightarrow \infty \) . For the approximated \(Q\) -function, we can apply the ergodic theorem since the smoothing distributions no longer depend on \(T\) .
| [1] | [
[
740,
743
]
] | https://openalex.org/W4250389103 |
5efa87c9-1ba8-410c-969d-fe3260382ba7 | In Assumption , we require the strong concavity of \(\bar{Q}(\cdot |\theta ^*)\) over the entire parameter space since the maximization step in our algorithm is global. Such a requirement could be avoided: if the maximization step is replaced by a gradient update (Gradient EM), then \(\bar{Q}(\cdot |\theta ^*)\) only needs to be strongly concave in a small region around \(\theta ^*\) . The price to pay is to assume knowledge on structural constants of \(\bar{Q}(\cdot |\theta ^*)\) (Lipschitz constant and strong concavity constant). See [1]} for an analysis of the gradient EM algorithm.
| [1] | [
[
545,
548
]
] | https://openalex.org/W2962737134 |
0ed42d85-3e80-4280-bd2d-dc7e3aeb5662 | The proof is given in Appendix REF , where we also show an upper bound on \(\gamma \) . The idea mirrors that of [1]} with problem-specific modifications. Algorithm can be regarded as a perturbed form of this population version algorithm, with convergence characterized in the following theorem.
| [1] | [
[
113,
116
]
] | https://openalex.org/W2962737134 |
34d4b257-3c00-4449-843d-583404f1a87e | The proof is provided in Appendix REF . Essentially, we use Theorem to show the uniform (in \(\theta \) and \(\mu \) ) strong consistency of \(M_{\mu ,T}(\theta ;\omega )\) as an estimator of \(\bar{M}(\theta )\) , following the standard analysis of M-estimators. A direct corollary of this argument is the high probability bound on the difference between \(M_{\mu ,T}(\theta ;\omega )\) and \(\bar{M}(\theta )\) , as shown in the first part of the theorem. Combining this high probability bound with Theorem REF and [1]} yields the final performance guarantee.
| [1] | [
[
522,
525
]
] | https://openalex.org/W2962737134 |
0419cf57-e1d1-4860-8111-19a2195c2fea | Appendix presents discussions that motivate Assumption REF . In particular, we show that Assumption REF approximately holds in a particular class of environment. Appendix provides details on Algorithm , including the comparison with the existing algorithm from [1]}, the forward-backward implementation and the derivation of the \(Q\) -function from (REF ). In Appendix , we prove our theoretical results from Section . Technical lemmas involved in the proofs are deferred to Appendix . Finally, Appendix presents details of our numerical example omitted from Section .
| [1] | [
[
264,
267
]
] | https://openalex.org/W2498991332 |
cfa1a178-4b46-4915-8604-f928e5d9603c | A Markov chain convergence result is restated in the following lemma, tailored to our need.
[[1]}, Theorem 4.3.16 restated]
With the Doeblin-type condition in (REF ), the Markov chain \(\lbrace X_t;\theta \rbrace _{t=1}^\infty \) with any \(\theta \in \) has a unique stationary distribution \(\nu _\theta \) . Moreover, for all \(\theta \in \) , \(x\in \mathcal {X}\) and \(t\in _+\) ,
\({Q^t_\theta (x,\cdot )-\nu _\theta }_\le (1-\delta )^{{t/2}}.\)
| [1] | [
[
93,
96
]
] | https://openalex.org/W4250389103 |
b2cb1693-c508-462b-ba4c-333270d53411 | First, we point out a technicality when comparing Algorithm to the algorithm from [1]}. The algorithm from [1]} learns a hierarchical policy following the standard options framework, not the options with failure framework considered in Algorithm . To draw direct comparison, we need to let \(\zeta =0\) in Algorithm . However, an error in the existing algorithm can be demonstrated without referring to \(\zeta \) .
| [1] | [
[
83,
86
],
[
108,
111
]
] | https://openalex.org/W2498991332 |
053074b3-012a-4fec-a76e-2cad0da795e6 | For simplicity, consider \(O_0\) fixed to \(o_0\in \) ; let \(2\le t\le T-1\) . Then, according to the definitions in [1]}, the (unnormalized) forward message is defined as
\(\alpha ^\theta _{t}(o_t,b_t)=¶_{\theta ,o_0,s_1}(A_{1:t}=a_{1:t},O_t=o_t,B_t=b_t|S_{2:t}=s_{2:t}).\)
| [1] | [
[
119,
122
]
] | https://openalex.org/W2498991332 |
e4a192c1-4adb-4cd9-99d1-172a3bd9d003 | We use the proportional symbol \(\propto \) to represent normalizing constants independent of \(o_t\) and \(b_t\) . [1]} claims that, for any \(o_t\) and \(b_t\) ,
\(\gamma ^{\theta }_{t|T}(o_t,b_t)\propto \alpha ^\theta _{t}(o_t,b_t)\beta ^\theta _{t|T}(o_t,b_t).\)
| [1] | [
[
118,
121
]
] | https://openalex.org/W2498991332 |
7c555f24-b294-49a8-8c91-e462500f4984 | For the claim in [1]} to be true, \(\mathbb {P}_{\theta ,o_0,s_1}(S_{t+1:T}=s_{t+1:T}|S_{t}=s_{t},A_{t}=a_{t},O_t=o_t,B_t=b_t)\) should not depend on \(o_t\) and \(b_t\) . Clearly this requirement does not hold in most cases, since the likelihood of the future observation sequence should depend on the currently applied option.
| [1] | [
[
17,
20
]
] | https://openalex.org/W2498991332 |
739a4549-e9eb-4bf8-97ef-b82f6a645adb | In our algorithm, as motivated by Section , we effectively consider the following joint distribution on the graphical model shown in Figure REF : the prior distribution of \((O_0,S_1)\) is \(\hat{\nu }\) , and the distribution of the rest of the graphical model is determined by an options with failure policy with parameters \(\zeta \) and \(\theta \) . From the EM literature [1]}, [2]}, the complete likelihood function is
\(L(s_{1:T},a_{1:T},o_{0:T},b_{1:T};\theta )=\hat{\nu }(o_0,s_1)\mathbb {P}_{\theta ,o_0,s_1}(S_{2:T}=s_{2:T},A_{1:T}=a_{1:T},O_{1:T}=o_{1:T},B_{1:T}=b_{1:T}).\)
| [1] | [
[
380,
383
]
] | https://openalex.org/W2962737134 |
8bfd937c-7571-4f26-81f4-556256598729 | where \(const\) is a constant independent of \(T\) and \(\omega \) .
The proof is provided in Appendix REF . Now we are ready to present the proof of Theorem REF step-by-step. The structure of this proof is similar to the standard analysis of HMM maximum likelihood estimators [1]}.
| [1] | [
[
281,
284
]
] | https://openalex.org/W4250389103 |
0b741f40-eaef-463f-bd67-44db8aee486e | 2. We need to prove the uniform (in \(\theta ,\theta ^{\prime }\in \) and \(\mu \in \mathcal {M}\) ) almost sure convergence of the \(Q\) -function \(Q_{\mu ,T}(\theta ^{\prime }|\theta ;\omega )\) to the population \(Q\) -function \(\bar{Q}(\theta ^{\prime }|\theta )\) . The proof is separated into three steps. First, we show the almost sure convergence of \(Q^s_{\infty ,T}(\theta ^{\prime }|\theta ;\omega )\) to \(\bar{Q}(\theta ^{\prime }|\theta )\) for all \(\theta ,\theta ^{\prime }\in \) using the ergodic theorem. Second, we extend this pointwise convergence to uniform (in \(\theta ,\theta ^{\prime }\) ) convergence using a version of the Arzelà-Ascoli theorem [1]}. Finally, from Lemma REF , the difference between \(Q_{\mu ,T}(\theta ^{\prime }|\theta ;\omega )\) and \(Q^s_{\infty ,T}(\theta ^{\prime }|\theta ;\omega )\) vanishes uniformly in \(\mu \) as \(T\rightarrow \infty \) .
| [1] | [
[
680,
683
]
] | https://openalex.org/W3144951960 |
b6c60e82-beea-43a5-8f99-87359946f1a6 | To extend the pointwise convergence in (REF ) to uniform (in \(\theta ,\theta ^{\prime }\) ) convergence, the following concept is required. The sequence \(\lbrace Q^s_{\infty ,T}(\theta ^{\prime }|\theta )\rbrace \) indexed by \(T\) as functions of \(\theta \) and \(\theta ^{\prime }\) is strongly stochastically equicontinuous [1]} if for any \(>0\) there exists \(\delta >0\) such that
\(\limsup _{T\rightarrow \infty }\sup _{\theta _1,\theta ^{\prime }_1,\theta _2,\theta ^{\prime }_2\in ;{\theta _1-\theta _2}_2+{\theta ^{\prime }_1-\theta ^{\prime }_2}_2\le \delta }\left|Q^s_{\infty ,T}(\theta ^{\prime }_1|\theta _1;\omega )-Q^s_{\infty ,T}(\theta ^{\prime }_2|\theta _2;\omega )\right|<,~P_{\theta ^*,\nu ^*}\text{-a.s.}\)
| [1] | [
[
334,
337
]
] | https://openalex.org/W3144951960 |
31051bfa-05c3-49aa-a3d8-b4c0d4607677 | [[1]}, Theorem 21.8 restated]
Given (REF ) and (REF ), as \(T\rightarrow \infty \) we have
\(\sup _{\theta ,\theta ^{\prime }\in }\left|Q^s_{\infty ,T}(\theta ^{\prime }|\theta ;\omega )-\bar{Q}(\theta ^{\prime }|\theta )\right|\rightarrow 0,~P_{\theta ^*,\nu ^*}\text{-a.s.}\)
| [1] | [
[
1,
4
]
] | https://openalex.org/W3144951960 |
acabf6b8-eb6b-44db-8d50-fd68ac791845 | 2. The proof of the second part mirrors the proof of [1]}. The main difference is the construction of the following self-consistency (a.k.a. fixed-point) condition.
[Self-consistency] With all the assumptions, \(\theta ^*=\bar{M}(\theta ^*)\) .
| [1] | [
[
53,
56
]
] | https://openalex.org/W2962737134 |
7bec1fd7-b743-440b-b293-2830c99b0527 | The proof of this lemma is presented in Appendix REF . Such a condition is used without proof in [1]} since it only considers i.i.d. samples, and the self-consistency condition for EM with i.i.d. samples is a well-known result. However, for the case of dependent samples like our graphical model, such a condition results from the stochastic convergence of the \(Q\) -function which is not immediate.
| [1] | [
[
97,
100
]
] | https://openalex.org/W2962737134 |