File size: 175,579 Bytes
8cf8144
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
1889
1890
1891
1892
1893
1894
1895
1896
1897
1898
1899
1900
1901
1902
1903
1904
1905
1906
1907
1908
1909
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919
1920
1921
1922
1923
1924
1925
1926
1927
1928
1929
1930
1931
1932
1933
1934
1935
1936
1937
1938
1939
1940
1941
1942
1943
1944
1945
1946
1947
1948
1949
1950
1951
1952
1953
1954
1955
1956
1957
1958
1959
1960
1961
1962
1963
1964
1965
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
2026
2027
2028
2029
2030
2031
2032
2033
2034
2035
2036
2037
2038
2039
2040
2041
2042
2043
2044
2045
2046
2047
2048
2049
2050
2051
2052
2053
2054
2055
2056
2057
2058
2059
2060
2061
2062
2063
2064
2065
2066
2067
2068
2069
2070
2071
2072
2073
2074
2075
2076
2077
2078
2079
2080
2081
2082
2083
2084
2085
2086
2087
2088
2089
2090
2091
2092
2093
2094
2095
2096
2097
2098
2099
2100
2101
2102
2103
2104
2105
2106
2107
2108
2109
2110
2111
2112
2113
2114
2115
2116
2117
2118
2119
2120
2121
2122
2123
2124
2125
2126
2127
2128
2129
2130
2131
2132
2133
2134
2135
2136
2137
2138
2139
2140
2141
2142
2143
2144
2145
2146
2147
2148
2149
2150
2151
2152
2153
2154
2155
2156
2157
2158
2159
2160
2161
2162
2163
2164
2165
2166
2167
2168
2169
2170
2171
2172
2173
2174
2175
2176
2177
2178
2179
2180
2181
2182
2183
2184
2185
2186
2187
2188
2189
2190
2191
2192
2193
2194
2195
2196
2197
2198
2199
2200
2201
2202
2203
2204
2205
2206
2207
2208
2209
2210
2211
2212
2213
2214
2215
2216
2217
2218
2219
2220
2221
2222
2223
2224
2225
2226
2227
2228
2229
2230
2231
2232
2233
2234
2235
2236
2237
2238
2239
2240
2241
2242
2243
2244
2245
2246
2247
2248
2249
2250
2251
2252
2253
2254
2255
2256
2257
2258
2259
2260
2261
2262
2263
2264
2265
2266
2267
2268
2269
2270
2271
2272
2273
2274
2275
2276
2277
2278
2279
2280
2281
2282
2283
2284
2285
2286
2287
2288
2289
2290
2291
2292
2293
2294
2295
2296
2297
2298
2299
2300
2301
2302
2303
2304
2305
2306
2307
2308
2309
2310
2311
2312
2313
2314
2315
2316
2317
2318
2319
2320
2321
2322
2323
2324
2325
2326
2327
2328
2329
2330
2331
2332
2333
2334
2335
2336
2337
2338
2339
2340
2341
2342
2343
2344
2345
2346
2347
2348
2349
2350
2351
2352
2353
2354
2355
2356
2357
2358
2359
2360
2361
2362
2363
2364
2365
2366
2367
2368
2369
2370
2371
2372
2373
2374
2375
2376
2377
2378
2379
2380
2381
2382
2383
2384
2385
2386
2387
2388
2389
2390
2391
2392
2393
2394
2395
2396
2397
2398
2399
2400
2401
2402
2403
2404
2405
2406
2407
2408
2409
2410
2411
2412
2413
2414
2415
2416
2417
2418
2419
2420
2421
2422
2423
2424
2425
2426
2427
2428
2429
2430
2431
2432
2433
2434
2435
2436
2437
2438
2439
2440
2441
2442
2443
2444
2445
2446
2447
2448
2449
2450
2451
2452
2453
2454
2455
2456
2457
2458
2459
2460
2461
2462
2463
2464
2465
2466
2467
2468
2469
2470
2471
2472
2473
2474
2475
2476
2477
2478
2479
2480
2481
2482
2483
2484
2485
2486
2487
2488
2489
2490
2491
2492
2493
2494
2495
2496
2497
2498
2499
2500
2501
2502
2503
2504
2505
2506
2507
2508
2509
2510
2511
2512
2513
2514
2515
2516
2517
2518
2519
2520
2521
2522
2523
2524
2525
2526
2527
2528
2529
2530
2531
2532
2533
2534
2535
2536
2537
2538
2539
2540
2541
2542
2543
2544
2545
2546
2547
2548
2549
2550
2551
2552
2553
2554
2555
2556
2557
2558
2559
2560
2561
2562
2563
2564
2565
2566
2567
2568
2569
2570
2571
2572
2573
2574
2575
2576
2577
2578
2579
2580
2581
2582
2583
2584
2585
2586
2587
2588
2589
2590
2591
2592
2593
2594
2595
2596
2597
2598
2599
2600
2601
2602
2603
2604
2605
2606
2607
2608
2609
2610
2611
2612
2613
2614
2615
2616
2617
2618
2619
2620
2621
2622
2623
2624
2625
2626
2627
2628
2629
2630
2631
2632
2633
2634
2635
2636
2637
2638
2639
2640
2641
2642
2643
2644
2645
2646
2647
2648
2649
2650
2651
2652
2653
2654
2655
2656
2657
2658
2659
2660
2661
2662
2663
2664
2665
2666
2667
2668
2669
2670
2671
2672
2673
2674
2675
2676
2677
2678
2679
2680
2681
2682
2683
2684
2685
2686
2687
2688
2689
2690
2691
2692
2693
2694
2695
2696
2697
2698
2699
2700
2701
2702
2703
2704
2705
2706
2707
2708
2709
2710
2711
2712
2713
2714
2715
2716
2717
2718
2719
2720
2721
2722
2723
2724
2725
2726
2727
2728
2729
2730
2731
2732
2733
2734
2735
2736
2737
2738
2739
2740
2741
2742
2743
2744
2745
2746
2747
2748
2749
2750
2751
2752
2753
2754
2755
2756
2757
2758
2759
2760
2761
2762
2763
2764
2765
2766
2767
2768
2769
2770
2771
2772
2773
2774
2775
2776
2777
2778
2779
2780
2781
2782
2783
2784
2785
2786
2787
2788
2789
2790
2791
2792
2793
2794
2795
2796
2797
2798
2799
2800
2801
2802
2803
2804
2805
2806
2807
2808
2809
2810
2811
2812
2813
2814
2815
2816
2817
2818
2819
2820
2821
2822
2823
2824
2825
2826
2827
2828
2829
2830
2831
2832
2833
2834
2835
2836
2837
2838
2839
2840
2841
2842
2843
2844
2845
2846
2847
2848
2849
2850
2851
2852
2853
2854
2855
2856
2857
2858
2859
2860
2861
2862
2863
2864
2865
2866
2867
2868
2869
2870
2871
2872
2873
2874
2875
2876
2877
2878
2879
2880
2881
2882
2883
2884
2885
2886
2887
2888
2889
2890
2891
2892
2893
2894
2895
2896
2897
2898
2899
2900
2901
2902
2903
2904
2905
2906
2907
2908
2909
2910
2911
2912
2913
2914
2915
2916
2917
2918
2919
2920
2921
2922
2923
2924
2925
2926
2927
2928
2929
2930
2931
2932
2933
2934
2935
2936
2937
2938
2939
2940
2941
2942
2943
2944
2945
2946
2947
2948
2949
2950
2951
2952
2953
2954
2955
2956
2957
2958
2959
2960
2961
2962
2963
2964
2965
2966
2967
2968
2969
2970
2971
2972
2973
2974
2975
2976
2977
2978
2979
2980
2981
2982
2983
2984
2985
2986
2987
2988
2989
2990
2991
2992
2993
2994
2995
2996
2997
2998
2999
3000
3001
3002
3003
3004
3005
3006
3007
3008
3009
3010
3011
3012
3013
3014
3015
3016
3017
3018
3019
3020
3021
3022
3023
3024
3025
3026
3027
3028
3029
3030
3031
3032
3033
3034
3035
3036
3037
3038
3039
3040
3041
3042
3043
3044
3045
3046
3047
3048
3049
3050
3051
3052
3053
3054
3055
3056
3057
3058
3059
3060
3061
3062
3063
3064
3065
3066
3067
3068
3069
3070
3071
3072
3073
3074
3075
3076
3077
3078
3079
3080
3081
3082
3083
3084
3085
3086
3087
3088
3089
3090
3091
3092
3093
3094
3095
3096
3097
3098
3099
3100
3101
3102
3103
3104
3105
3106
3107
3108
3109
3110
3111
3112
3113
3114
3115
3116
3117
3118
3119
3120
3121
3122
3123
3124
3125
3126
3127
3128
3129
3130
3131
3132
3133
3134
3135
3136
3137
3138
3139
3140
3141
3142
3143
3144
3145
3146
3147
3148
3149
3150
3151
3152
3153
3154
3155
3156
3157
3158
3159
3160
3161
3162
3163
3164
3165
3166
3167
3168
3169
3170
3171
3172
3173
3174
3175
3176
3177
3178
3179
3180
3181
3182
3183
3184
3185
3186
3187
3188
3189
3190
3191
3192
3193
3194
3195
3196
3197
3198
3199
3200
3201
3202
3203
3204
3205
3206
3207
3208
3209
3210
3211
3212
3213
3214
3215
3216
3217
3218
3219
3220
3221
3222
3223
3224
3225
3226
3227
3228
3229
3230
3231
3232
3233
3234
3235
3236
3237
3238
3239
3240
3241
3242
3243
3244
3245
3246
3247
3248
3249
3250
3251
3252
3253
3254
3255
3256
3257
3258
3259
3260
3261
3262
3263
3264
3265
3266
3267
3268
3269
3270
3271
3272
3273
3274
3275
3276
3277
3278
3279
3280
3281
3282
3283
3284
3285
3286
3287
3288
3289
3290
3291
3292
3293
3294
3295
3296
3297
3298
3299
3300
3301
3302
3303
3304
3305
3306
3307
3308
3309
3310
3311
3312
3313
3314
3315
3316
3317
3318
3319
3320
3321
3322
3323
3324
3325
3326
3327
3328
3329
3330
3331
3332
3333
3334
3335
3336
3337
3338
3339
3340
3341
3342
3343
3344
3345
3346
3347
3348
3349
3350
3351
3352
3353
3354
3355
3356
3357
3358
3359
3360
3361
3362
3363
3364
3365
3366
3367
3368
3369
3370
3371
3372
3373
3374
3375
3376
3377
3378
3379
3380
3381
3382
3383
3384
3385
3386
3387
3388
3389
3390
3391
3392
3393
3394
3395
3396
3397
3398
3399
3400
3401
3402
3403
3404
3405
3406
3407
3408
3409
3410
3411
3412
3413
3414
3415
3416
3417
3418
3419
3420
3421
3422
3423
3424
3425
3426
3427
3428
3429
3430
3431
3432
3433
3434
3435
3436
3437
3438
3439
3440
3441
3442
3443
3444
3445
3446
3447
3448
3449
3450
3451
3452
3453
3454
3455
3456
3457
3458
3459
3460
3461
3462
3463
3464
3465
3466
3467
3468
3469
3470
3471
3472
3473
3474
3475
3476
3477
3478
3479
3480
3481
3482
3483
3484
3485
3486
3487
3488
3489
3490
3491
3492
3493
3494
3495
3496
3497
3498
3499
3500
3501
3502
3503
3504
3505
3506
3507
3508
3509
3510
3511
3512
3513
3514
3515
3516
3517
3518
3519
3520
3521
3522


\documentclass[orivec]{llncs}
\bibliographystyle{splncs}
\let\proof\relax
\let\endproof\relax
\usepackage{amsmath,amssymb,amsthm,bbm}
\usepackage{mathpartir}
\usepackage{stmaryrd}
\usepackage{soul}
\usepackage{listings}
\usepackage{proof}
\usepackage{color}
\usepackage{hyperref}
\usepackage{float}
\usepackage{rotating}
\usepackage{mdframed}
\usepackage{enumitem}
\usepackage{xifthen}
\usepackage{tikz}
\usepackage{multirow}
\usepackage[misc]{ifsym}


\addtolength{\belowcaptionskip}{-0.5em}


\newif\ifdraft
\draftfalse


\newcommand{\esyntax}[1]{\mathsf{#1}}


\newcommand{\rname}[1]{[\textsf{#1}]}
\newcommand{\jrhol}[7]{#1 \mid #2 \vdash #3 : #4 \sim #5 : #6 \mid #7}
\newcommand{\jrholu}[5]{#1 \mid #2 \vdash #3 \sim #4 \mid #5}
\newcommand{\sjrhol}[7]{#2 \vdash #3 \sim #5 \mid #7}
\newcommand{\juhol}[5]{#1 \mid #2 \vdash #3 : #4 \mid #5}
\newcommand{\sjuhol}[5]{#2 \vdash #3 \mid #5}
\newcommand{\jhol}[3]{#1 \mid #2 \vdash #3}
\newcommand{\vhol}[3]{#1;#2 \models #3}
\newcommand{\jholn}[3]{#1 \mid #2 \vdash_{\sf HOL} #3}
\newcommand{\jlc}[3]{#1 \vdash #2 : #3}
\newcommand{\jwf}[2]{#1 \vdash #2 \; \sf{wf}}
\newcommand{\subst}[2]{[#2/ #1]}
\newcommand{\ltag}{_1}
\newcommand{\rtag}{_2}
\newcommand{\res}{\mathbf{r}}
\newcommand{\sem}[1]{\llbracket #1 \rrbracket}
\newcommand{\semv}[2]{\llbracket #1 \rrbracket_{#2} }
\newcommand{\semexp}[1]{\llparenthesis #1 \rrparenthesis}
\newcommand{\semexpv}[2]{\llparenthesis #1 \rrparenthesis_{#2} }

\newcommand{\pair}[2]{\langle #1, #2 \rangle}
\newcommand{\nat}{\mathbb{N}}
\newcommand{\zint}{\mathbb{Z}}
\newcommand{\real}{\mathbb{R}}
\newcommand{\unit}{\mathbb{U}}
\newcommand{\listt}[1]{\mathsf{list}_{#1}}
\newcommand{\bool}{\mathbb{B}}
\newcommand{\tbool}{\mathsf{tt}}
\newcommand{\fbool}{\mathsf{ff}}
\newcommand{\nil}{[]}
\newcommand{\cons}[2]{#1\! ::\! #2}
\newcommand{\casebool}[3]{{\rm case}\ #1\ {\sf of}\ \tbool \mapsto #2 ; \fbool \mapsto #3}
\newcommand{\casenat}[3]{{\rm case}\ #1\ {\sf of}\ 0 \mapsto #2 ; S \mapsto #3}
\newcommand{\caselist}[3]{{\rm case}\ #1\ {\rm of}\ \nil \mapsto #2 ; \cons{\_}{\_} \mapsto #3}  
\newcommand{\defsubst}[2]{\subst{\res\ltag}{#1}\subst{\res\rtag}{#2}}
\newcommand{\wdef}[3]{\mathcal{D}\hspace{-0.2mm}\mathit{ef}(#1,#2,#3)}
\newcommand{\letrec}[3]{{\rm letrec}\ {\mathit #1}\ #2 = #3}
\newcommand{\assn}{\star}
\newcommand{\cost}[1]{\llparenthesis #1 \rrparenthesis}
\newcommand{\costt}[1]{\lfloor #1 \rfloor}
\newcommand{\rcost}[1]{\| #1 \|}
\newcommand{\rcostt}[1]{\llfloor #1 \rrfloor}
\newcommand{\defeq}{\triangleq}
\newcommand{\erase}[1]{| #1 |}

\newcommand{\reftype}[3]{\ifthenelse{\isempty{#1}}{\{\res : #2 \mid #3\}}{\{#1 : #2 \mid #3\}}}
\newcommand{\ertype}[2]{\ifthenelse{\isempty{#2}}{\lfloor #1 \rfloor}{\lfloor #1 \rfloor(#2)}}
\newcommand{\errtype}[2]{\ifthenelse{\isempty{#2}}{\llfloor #1 \rrfloor}{\llfloor #1 \rrfloor(#2)}}
\newcommand{\erfor}[1]{| #1 |}
\newcommand{\embind}[1]{[#1]}
\newcommand{\emfor}[2]{\ifthenelse{\isempty{#2}}{#1}{#1\subst{\res}{#2}}}
\newcommand{\emterm}[2]{\ifthenelse{\isempty{#2}}{#1}{#1\subst{\res}{#2}}}
\newcommand{\emtyj}[3]{\juhol{\erfor{#1}}{\ertype{#1}{}}{#2}{\erfor{#3}}{\ertype{#3}{\res}}}
\newcommand{\jreft}[3]{#1 \vdash #2 : #3}
\newcommand{\pitype}[3]{\Pi(#1:#2).#3}

\newcommand{\ttmonad}[1]{\mathbb{T}_{#1}}
\newcommand{\eret}[1]{\eta_{#1}}
\newcommand{\ebind}{\mathsf{bind}}
\newcommand{\dccform}[2]{\lfloor #2 \rfloor_{#1}}
\newcommand{\adversary}{a}
\newcommand{\boolform}[1]{R^{\bool}_{#1}}

\newcommand{\wexec}{\mbox{\scriptsize exec}}
\newcommand{\wdiff}{\mbox{\scriptsize diff}}
\newcommand{\ldiff}{\lesssim}
\newcommand{\uarr}[2]{\mathrel{\xrightarrow[]{\wexec(#1,#2)}}} 
\newcommand{\tforall}[3]{\forall#1\overset{\wexec(#2,#3)}{::}S.\,}
\newcommand{\texists}[1]{\exists#1{::}S.\,}
\newcommand{\tarrd}[1]{\mathrel{\xrightarrow{\wdiff(#1)}}}
\newcommand{\tforalld}[2]{\forall#1\overset{\wdiff(#2)}{::}S.\,}
\newcommand{\tbox}[1]{\square\,#1}
\newcommand{\jtype}[4]{\mathrel{\vdash_{#1}^{#2} {#3} : #4}}
\newcommand{\jtypediff}[4]{\mathrel{\vdash {#2} \ominus {#3} \ldiff #1 : {#4}}}
\newcommand{\econs}{\mbox{cons}}
\newcommand{\ulist}[2]{\mbox{list}_{#2}[#1]}
\newcommand{\ctx}{\Delta; \Phi_a; \Gamma}
\newcommand{\kw}[1]{\mathtt{#1}}
\newcommand{\econst}{\kw{n}}
\newcommand{\trint}{\mbox{int}_r}
\newcommand{\tlist}[3]{\mbox{list}[#1]^{#2}\,#3}
\newcommand{\ecase}{\mbox{\;case\;}} 
\newcommand{\eof}{\mbox{\;of\;}}
\newcommand{\eas}{\mbox{\;as\;}} 
\newcommand{\ecelim}{\mbox{celim\;}}
\newcommand{\enil}{\mbox{nil\;}} 
\newcommand{\fiv}[1]{\text{FIV}(#1)}
\newcommand{\eLam}{ \Lambda}
\newcommand{\elam}{ \lambda} 
\newcommand{\eApp}{ [\,]\,}
\newcommand{\sat}[1]{\models#1}
\newcommand{\jsubtype}[2]{\sat#1\sqsubseteq#2}
\newcommand{\trmo}[1]{|#1|}
\newcommand{\grt}{A}
\newcommand{\tcho}[1]{U\,#1} 
\newcommand{\octx}{\Delta; \Phi_a; \Omega}
\newcommand{\tint}{\mbox{int}}
\newcommand{\kapp}{c_{app}}


\def\tyle{\preceq}

\ifdraft
\newcommand{\aag}[1]{\textit{\color{blue}[AA]: #1}}
\newcommand{\gb}[1]{\textit{\color{red}[GB]: #1}}
\newcommand{\mg}[1]{\textit{\color{cyan}[MG]: #1}}
\newcommand{\dg}[1]{\textit{\color{brown}[DG]: #1}}
\newcommand{\ales}[1]{\textit{\color{green}[AB]: #1}}
\newcommand{\lb}[1]{\textit{\color{green}[LB]: #1}}
\else
\newcommand{\aag}[1]{}
\newcommand{\gb}[1]{}
\newcommand{\mg}[1]{}
\newcommand{\dg}[1]{}
\newcommand{\ales}[1]{}
\newcommand{\lb}[1]{}
\fi

\theoremstyle{definition}
\newtheorem{thm}{Theorem}
\newtheorem{lem}[thm]{Lemma}
\newtheorem{cor}[thm]{Corollary}
\newtheorem{prop}[thm]{Proposition}

\newtheorem{eg}[thm]{Example}
\newtheorem{hypothesis}[thm]{Hypothesis}
\newtheorem{motivation}{Motivation}
\newtheorem{defn}[thm]{Definition}

\newcommand{\RHOL}{\mathsf{2HOL}}

\newcommand{\renv}[1]{\mathcal{#1}}
\newcommand{\vbar}[0]{\mathrel{|}}
\newcommand{\rtprod}[3]{\Pi ({#1} :: {#2}) .\, {#3}}
\newcommand{\rtref}[3]{\{ {#1} :: {#2} \vbar {#3} \}}
\newcommand{\tprod}[3]{\Pi ({#1} : {#2}) .\, {#3}}
\newcommand{\tref}[3]{\{ {#1} : {#2} \vbar {#3} \}}

\newcommand{\tyinterp}[1]{\llbracket {#1} \rrbracket}
\newcommand{\interp}[2]{\llbracket {#2} \rrbracket_{#1}}
\newcommand{\rinterp}[2]{\llparenthesis {#2} \rrparenthesis_{#1}}

\DeclareDocumentCommand{\later}{ o m}{
  \IfNoValueTF{#1}
  {\triangleright #2}
  {\triangleright\hrt{#1} . #2}}
\DeclareDocumentCommand{\latern}{ o m}{
  \IfNoValueTF{#1}
  {\triangleright #2}
  {\triangleright #1 . #2}}
\newcommand{\vrt}[1]{\left[\begin{array}{l} #1 \end{array}\right]}
\newcommand{\hrt}[1]{\left[ #1 \right]}
\newcommand{\dsubst}[3]{\triangleright[#1 \leftarrow #2].#3}
\newcommand{\nextt}[2]{\triangleright[#1].#2}
\newcommand{\prev}[1]{{\rm prev}\ #1}
\newcommand{\boxx}[1]{{\rm box}\ #1}
\newcommand{\jguhol}[7]{#1 \mid #2 \mid #3 \mid #4 \vdash #5 : #6 \mid #7}
\newcommand{\jguholsc}[7]{\Omega \vdash #5 : #6 \mid #7}
\newcommand{\letbox}[3]{{\rm letb}\ #1 \leftarrow #2\ {\rm in}\ #3}
\newcommand{\unbox}[1]{{\rm unbox}\ #1}
\newcommand{\letconst}[3]{{\rm letc}\ #1 \leftarrow #2\ {\rm in}\ #3}
\newcommand{\conshat}[2]{#1\hat{::}#2}
\newcommand{\fix}[2]{{\rm fix}\ #1.\ #2}
\newcommand{\rec}[2]{{\rm rec}\ #1.\ #2}
\newcommand{\app}{\circledast}
\newcommand{\jgrhol}[9]{#1 \mid #2 \mid #3 \mid #4 \vdash #5 : #6 \sim #7 : #8 \mid #9}
\newcommand{\jgrholsc}[9]{\Omega \vdash #5 : #6 \sim #7 : #8 \mid #9}
\newcommand{\jgrholnoc}[5]{\vdash #1 : #2 \sim #3 : #4 \mid #5}
\newcommand{\jgrholnocnot}[3]{\vdash #1 \sim #2 \mid #3}
\newcommand{\jghol}[5]{#1 \mid #2 \mid #3 \mid #4 \vdash #5}
\newcommand{\jgholsc}[5]{\Omega \vdash #5}
\newcommand{\conj}{\mathrel{\wedge}}
\newcommand{\trees}{\mathcal{S}}
\newcommand{\ot}{\leftarrow}
\newcommand{\xot}[1]{\xleftarrow{#1}}
\newcommand{\xto}[1]{\xrightarrow{#1}}
\newcommand{\To}{\Rightarrow}
\newcommand{\oT}{\Leftarrow}
\newcommand{\ii}{\mathit}
\newcommand{\thesystem}{gRHOL}
\newcommand{\inj}[1]{{\rm inj}_{#1}}
\newcommand{\CASE}{\operatorname{\mathrm{case}}}
\newcommand{\OF}{\operatorname{\mathrm{of}}}
\newcommand{\IN}{\operatorname{\mathrm{in}}}
\newcommand{\hd}{\operatorname{\mathrm{hd}}}
\newcommand{\tl}{\operatorname{\mathrm{tl}}}


\newcommand{\Distr}{\mathsf{D}}
\newcommand{\Str}[1]{\ifstrempty{#1}{\operatorname{\mathrm{Str}}}{\operatorname{\mathrm{Str}}_{#1}}}
\newcommand{\munit}[1]{\mathsf{munit}(#1)}
\newcommand{\mlet}[3]{\mathsf{let}~#1=#2~\mathsf{in}~#3}
\newcommand{\carac}[1]{\mathbbm{1}_{#1}}
\newcommand{\supp}{\mathsf{supp}}

\newcommand{\conslist}[2]{#1 :: #2}
\newcommand{\emptylist}{\operatorname{[]}}

\newcommand{\fold}{\operatorname{\mathsf{fold}}}
\newcommand{\unfold}{\operatorname{\mathsf{unfold}}}
\newcommand{\markov}{\operatorname{\mathsf{markov}}}

\newcommand{\isetsep}{\;\ifnum\currentgrouptype=16 \middle\fi|\;}

\newcommand{\bern}[1]{\mathcal{B}(#1)}
\newcommand{\unif}[1]{\mathcal{U}_{#1}}
\newcommand{\rfrac}[2]{{}^{#1}\!/\!_{#2}}
\newcommand{\All}{\operatorname{All}}
\newcommand{\coupling}[3]{#2 \mathrel{\mathcal{L}\left(#1\right)} #3}
\newcommand{\coupl}[3]{\diamond_{#1, #2}. #3}
\newcommand{\sP}[2]{\operatorname{Prob}_{#1}[#2]}
\newcommand{\shrinkcaption}{\vspace{-1em}}




 
\begin{document}

\title{Relational Reasoning for Markov Chains in a Probabilistic Guarded Lambda Calculus}
\author{Alejandro Aguirre\inst{1}$^\text{(\Letter)}$
        \and Gilles Barthe\inst{1}
        \and Lars Birkedal\inst{2}
        \and Ale\u s Bizjak\inst{2}
        \and \\ Marco Gaboardi\inst{3}
        \and Deepak Garg\inst{4}}
\institute{IMDEA Software Institute
           \\ \texttt{alejandro.aguirre@imdea.org}
           \and Aarhus University
           \and University at Buffalo, SUNY
           \and MPI-SWS}




\maketitle

\begin{abstract}
We extend the simply-typed guarded $\lambda$-calculus with discrete
probabilities and endow it with a program logic for reasoning about
relational properties of guarded probabilistic computations.  This
provides a framework for programming and reasoning about infinite
stochastic processes like Markov chains.  We demonstrate the logic
sound by interpreting its judgements in the topos of trees and by
using probabilistic couplings for the semantics of relational
assertions over distributions on discrete types.

The program logic is designed to support syntax-directed proofs in the
style of relational refinement types, but retains the expressiveness
of higher-order logic extended with discrete distributions, and the
ability to reason relationally about expressions that have different
types or syntactic structure. In addition, our proof system leverages
a well-known theorem from the coupling literature to justify better
proof rules for relational reasoning about probabilistic
expressions. We illustrate these benefits with a broad range of
examples that were beyond the scope of previous systems, including
shift couplings and lump couplings between random walks.
\end{abstract}



\section{Introduction}

Stochastic processes are often used in mathematics, physics, biology
or finance to model evolution of systems with uncertainty. In
particular, Markov chains are ``memoryless'' stochastic processes, in
the sense that the evolution of the system depends only on the current
state and not on its history. Perhaps the most emblematic example of a
(discrete time) Markov chain is the simple random walk over the
integers, that starts at 0, and that on each step moves one position
either left or right with uniform probability.  Let $p_i$ be the
position at time $i$. Then, this Markov chain can be described as:
\[
     p_0 = 0 \quad\quad
     p_{i+1} = \begin{cases}
       p_i + 1\ \text{with probability}\ 1/2 \\
       p_i - 1\ \text{with probability}\ 1/2
     \end{cases}
\]

The goal of this paper is to develop a programming and reasoning
framework for probabilistic computations over infinite objects, such
as Markov chains. Although programming and reasoning frameworks for
infinite objects and probabilistic computations are well-understood in
isolation, their combination is challenging. In particular, one must
develop a proof system that is powerful enough for proving interesting
properties of probabilistic computations over infinite objects, and
practical enough to support effective verification of these
properties.


\paragraph*{Modelling probabilistic infinite objects}
A first challenge is to model probabilistic infinite objects. We focus
on the case of Markov chains, due to its importance. A (discrete-time)
Markov chain is a sequence of random variables $\{X_i\}$ over some
fixed type $T$ satisfying some independence property. Thus, the
straightforward way of modelling a Markov chain is as a \emph{stream
  of distributions} over $T$. Going back to the simple example
outlined above, it is natural to think about this kind of
\emph{discrete-time} Markov chain as characterized by the sequence of
positions $\{p_i\}_{i \in \nat}$, which in turn can be described as an
infinite set indexed by the natural numbers. This suggests that a
natural way to model such a Markov chain is to use \emph{streams} in
which each element is produced \emph{probabilistically} from the
previous one. However, there are some downsides to this
representation. First of all, it requires explicit reasoning about
probabilistic dependency, since $X_{i+1}$ depends on $X_i$. Also, we
might be interested in global properties of the executions of the
Markov chain, such as ``The probability of passing through the initial
state infinitely many times is 1''. These properties are naturally
expressed as properties of the whole stream.  For these reasons, we
want to represent Markov chains as \emph{distributions over
  streams}. Seemingly, one downside of this representation is that the
set of streams is not countable, which suggests the need for
introducing heavy measure-theoretic machinery in the semantics of the
programming language, even when the underlying type is discrete or
finite.

Fortunately, measure-theoretic machinery can be avoided
(for discrete distributions) by developing a probabilistic extension
of the simply-typed guarded $\lambda$-calculus and giving a semantic
interpretation in the topos of trees~\cite{CBGB16}. Informally, the
simply-typed guarded $\lambda$-calculus~\cite{CBGB16} extends the
simply-typed lambda calculus with a \emph{later} modality, denoted by
$\later$. The type $\later{A}$ ascribes expressions that are available
one unit of logical time in the future. The $\later$ modality allows
one to model infinite types by using \lq\lq
finite\rq\rq\ approximations. For example, a stream of natural numbers
is represented by the sequence of its (increasing) prefixes in the
topos of trees. The prefix containing the first $i$ elements has the
type $S_i \defeq \nat \times \later{\nat} \times \ldots \times
\later^{(i-1)}{\nat}$, representing that the first element is
available now, the second element a unit time in the future, and so
on. This is the key to representing probability distributions over
infinite objects without measure-theoretic semantics: We model
probability distributions over non-discrete sets as discrete
distributions over their (the sets') approximations. For example, a
distribution over streams of natural numbers (which a priori would
be non-discrete since the set of streams is uncountable) would be
modelled by a \emph{sequence of distributions} over the finite
approximations $S_1, S_2, \ldots$ of streams. Importantly, since each
$S_i$ is countable, each of these distributions can be
discrete.


\dg{As Alejandro also noted after his example that I commented out,
  it's unclear what's so specific or special about the $\later$
  modality. It seems that we could use \emph{any} approximating
  sequence of discrete sets to avoid the issue with
  measure-theory. What's special about $\later$ and guards in our
  context?}
\lb{the use of $\later$ and guards just allow us to use types to
  ensure productivity (and contractiveness to ensure existence of
  fixed points)}
\mg{My understanding is that there is nothing special but we need in the logic to have a way to talk about these approximations, and
  $\later$ and guards gives us a natural way to do this. no?}

















\paragraph*{Reasoning about probabilistic computations}
Probabilistic computations exhibit a rich set of properties. One
natural class of properties is related to probabilities of events,
saying, for instance, that the probability of some event $E$ (or of an
indexed family of events) increases at every iteration. However,
several interesting properties of probabilistic computation, such as
stochastic dominance or convergence (defined below) are relational, in
the sense that they refer to two runs of two processes. In principle,
both classes of properties can be proved using a higher-order logic
for probabilistic expressions, e.g.\, the internal logic of the topos
of trees, suitably extended with an axiomatization of finite
distributions. However, we contend that an alternative approach
inspired from refinement types is desirable and provides better
support for effective verification. More specifically, reasoning in a
higher-order logic, e.g.\, in the internal logic of the topos of trees,
does not exploit the \emph{structure of programs} for non-relational
reasoning, nor the \emph{structural similarities} between programs for
relational reasoning. As a consequence, reasoning is more involved.
To address this issue, we define a relational proof system that
exploits the structure of the expressions and supports syntax-directed
proofs, with necessary provisions for escaping the syntax-directed
discipline when the expressions do not have the same structure. The
proof system manipulates judgements of the form:
\begin{equation*}
\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi}
\label{eq:judgement-2}
\end{equation*}
where $\Delta$ and $\Gamma$ are two typing contexts, $\Sigma$ and
$\Psi$ respectively denote sets of assertions over variables in these
two contexts, $t_1$ and $t_2$ are well-typed expressions of type $A_1$
and $A_2$, and $\phi$ is an assertion that may contain the special
variables $\res\ltag$ and $\res\rtag$ that respectively correspond to
the values of $t_1$ and $t_2$. The context $\Delta$ and $\Gamma$, the
terms $t_1$ and $t_2$ and the types $A_1$ and $A_2$ provide a
specification, while $\Sigma$, $\Psi$, and $\phi$ are useful for
reasoning about relational properties over $t_1,t_2$, their inputs and
their outputs. This form of judgement is similar to that of Relational
Higher-Order Logic \cite{ABGGS17}, from which our system draws inspiration.

In more detail, our relational logic comes with typing rules that
allow one to reason about relational properties by exploiting as much
as possible the syntactic similarities between $t_1$ and $t_2$, and to
fall back on pure logical reasoning when these are not available. In
order to apply relational reasoning to guarded computations the logic
provides relational rules for the later modality $\later{}$ and for a
related modality $\square{}$, called ``constant''. These rules allow
the relational verification of general relational properties that go
beyond the traditional notion of program equivalence and, moreover,
they allow the verification of properties of guarded computations over
different types. The ability to reason about computations of different
types provides significant benefits over alternative formalisms for
relational reasoning. For example, it enables reasoning about
relations between programs working on different data structures,
e.g. a relation between a program working on a stream of natural
numbers, and a program working on a stream of pairs of natural
numbers, or having different structures, e.g. a relation between
an application and a case expression.


Importantly, our approach for reasoning formally about probabilistic
computations is based on \emph{probabilistic couplings}, a standard
tool from the analysis of Markov chains~\cite{Lindvall02,Thorisson00}.
From a verification perspective, probabilistic couplings go beyond
equivalence properties of probabilistic programs, which have been
studied extensively in the verification literature, and yet support
compositional reasoning~\cite{BartheEGHSS15,BartheGHS17}. The main
attractive feature of coupling-based reasoning is that it limits the
need of explicitly reasoning about the probabilities---this avoids
complex verification conditions. We provide sound proof rules for
reasoning about probabilistic couplings. Our rules make several
improvements over prior relational verification logics based on
couplings. First, we support reasoning over probabilistic processes of
different types. Second, we use Strassen's
theorem~\cite{strassen1965existence} a remarkable result about
probabilistic couplings, to achieve greater expressivity. Previous
systems required to prove a bijection between the sampling spaces to
show the existence of a coupling~\cite{BartheEGHSS15,BartheGHS17},
Strassen's theorem gives a way to show their existence which is
applicable in settings where the bijection-based approach cannot be
applied.  And third, we support reasoning with what are called shift
couplings, coupling which permits to relate the states of two Markov
chains at possibly different times (more explanations below).





\paragraph*{Case studies}
We show the flexibility of our formalism by verifying several examples
of relational properties of probabilistic computations, and Markov
chains in particular. These examples cannot be verified with existing
approaches.

First, we verify a classic example of probabilistic non-interference
which requires the reasoning about computations at different
types. Second, in the context of Markov chains, we verify an example
about stochastic dominance which exercises our more general rule for
proving the existence of couplings modelled by expressions of
different types. Finally, we verify an example involving shift
relations in an infinite computation. This style of reasoning is
motivated by \lq\lq shift\rq\rq\ couplings in Markov chains. In
contrast to a standard coupling, which relates the states of two
Markov chains at the same time $t$, a shift coupling relates the
states of two Markov chains at possibly different times. Our specific
example relates a standard random walk (described earlier) to a
variant called a lazy random walk; the verification requires relating
the state of standard random walk at time $t$ to the state of the lazy
random walk at time $2t$. We note that this kind of reasoning is
impossible with conventional relational proof rules even in a
non-probabilistic setting. Therefore, we provide a novel family of
proof rules for reasoning about shift relations. At a high level, the
rules combine a careful treatment of the later and constant modalities
with a refined treatment of fixpoint operators, allowing us to relate
different iterates of function bodies.


\subsection*{Summary of contributions}
With the aim of providing a general framework for programming and
reasoning about Markov chains, the three main contributions of this work are:
\begin{enumerate}
  \item A probabilistic extension of the guarded $\lambda$-calculus,
    that enables the definition of Markov chains as discrete
    probability distributions over streams.
  \item A relational logic based on coupling to reason in a
    syntax-directed manner about (relational) properties of Markov
    chains. This logic supports reasoning about programs that have
    different types and structures. Additionally, this logic uses results from the
    coupling literature to achieve greater expressivity than previous
    systems.
  \item An extension of the relational logic that allows to relate the
    states of two streams at possibly different times. This extension
    supports reasoning principles, such as shift couplings, that escape
    conventional relational logics.
  \end{enumerate}
 






















 
\section{Mathematical preliminaries}
This section reviews the definition of discrete probability
sub-distributions and introduces mathematical couplings.
\begin{definition}[Discrete probability distribution]
Let $C$ be a discrete (i.e., finite or countable) set. A (total)
distribution over $C$ is a function $\mu : C \to [0,1]$ such that $
\sum_{x\in C} \mu(x) = 1 .$ The support of a distribution $\mu$ is the
set of points with non-zero probability, $ \supp\ \mu \defeq \{x \in C
\mid \mu(x) > 0 \} .$ We denote the set of distributions over $C$ as
$\Distr(C)$. Given a subset $E \subseteq C$, the probability of
sampling from $\mu$ a point in $E$ is denoted $\Pr_{x\leftarrow \mu}[x
  \in E]$, and is equal to $\sum_{x \in E} \mu(x)$.
\end{definition}


\begin{definition}[Marginals]
Let $\mu$ be a distribution over a product space $C_1\times C_2$. The
first (second) marginal of $\mu$ is another distribution
$\Distr(\pi_1)(\mu)$ ($\Distr(\pi_2)(\mu)$) over $C_1$ ($C_2$) defined
as:
\[\Distr(\pi_1)(\mu)(x) = \sum_{y \in C_2} \mu(x,y) \qquad
  \left(\Distr(\pi_2)(\mu)(y) = \sum_{x \in C_1} \mu(x,y) \right)
\]
\end{definition}

\subsubsection{Probabilistic couplings}
Probabilistic couplings are a fundamental tool in the analysis of
Markov chains. When analyzing
a relation between two probability distributions it is sometimes useful to
consider instead a distribution over the product space that somehow ``couples''
the randomness in a convenient manner.

Consider for instance the case of the following Markov chain, which
counts the total amount of tails observed when
tossing repeatedly a biased coin with probability of tails $p$:
\[
     n_0 = 0 \quad\quad
     n_{i+1} = \left\{\begin{array}{l}
       n_i + 1\ \text{with probability}\ p \\
       n_i\ \text{with probability}\ (1-p)
     \end{array}\right.
\]
If we have two biased coins with probabilities of tails $p$ and $q$ with $p\leq q$ and we
respectively observe $\{n_i\}$ and $\{m_i\}$ we would expect that, in
some sense, $n_i \leq m_i$ should hold for all $i$ (this property is
known as stochastic dominance).  A formal proof of this fact using
elementary tools from probability theory would require to compute the
cumulative distribution functions for $n_i$ and $m_i$ and then to
compare them. The coupling method reduces this proof to showing a way
to pair the coin flips so that if the first coin shows tails, so does
the second coin.



We now review the definition of couplings and state relevant
properties.
\begin{definition}[Couplings]
Let $\mu_1\in\Distr(C_1)$ and $\mu_2\in\Distr(C_2)$, and 
$R\subseteq C_1\times C_2$.
\begin{itemize}
\item A distribution $\mu\in\Distr(C_1\times C_2)$ is a coupling for
  $\mu_1$ and $\mu_2$ iff its first and second marginals coincide with
  $\mu_1$ and $\mu_2$ respectively, i.e.\, $\Distr(\pi_1)(\mu)=\mu_1$
  and $\Distr(\pi_2)(\mu)=\mu_2$.

\item A distribution $\mu\in\Distr(C_1\times C_2)$ is a $R$-coupling
  for $\mu_1$ and $\mu_2$ if it is a coupling for $\mu_1$ and $\mu_2$
  and, moreover, $\Pr_{(x_1,x_2)\leftarrow \mu} [R~x_1~x_2]=1$, i.e., if
  the support of the distribution $\mu$ is included in $R$.\end{itemize}
Moreover, we write $\coupl{\mu_1}{\mu_2}{R}$ iff there exists a
$R$-coupling for $\mu_1$ and $\mu_2$.


\end{definition}
Couplings always exist. For instance, the product distribution of two
distributions is always a coupling.
Going back to the example about the two coins, it can be proven by computation
that the following is a coupling that lifts the less-or-equal relation
($0$ indicating heads and $1$ indicating tails):
\[\left\{\begin{array}{lll}
   &(0,0) \ \text{w/ prob}\ (1-q)\quad\quad
   &(0,1) \ \text{w/ prob}\ (q-p) \\
   &(1,0) \ \text{w/ prob}\ 0\quad\quad
   &(1,1) \ \text{w/ prob}\ p
\end{array}\right.\]

The following theorem in~\cite{strassen1965existence} gives a
necessary and sufficient condition for the existence of $R$-couplings
between two distributions. The theorem is remarkable in the sense that
it proves an equivalence between an existential property (namely the
existence of a particular coupling) and a universal property
(checking, for each event, an inequality between probabilities).
\begin{theorem}[Strassen's theorem]
  Consider $\mu_1\in\Distr(C_1)$ and $\mu_2\in\Distr(C_2)$, and $R\subseteq
  C_1 \times C_2$. Then $\coupl{\mu_1}{\mu_2}{R}$ iff for every
  $X \subseteq C_1$, $\Pr_{x_1\leftarrow \mu_1}[x_1\in X] \leq
  \Pr_{x_2\leftarrow \mu_2}[x_2\in R(X)]$, where $R(X)$ is the image
  of $X$ under $R$, i.e.\, $R(X) =\{ y \in C_2 \mid \exists x \in
  X.~R~x~y\}$.
\end{theorem}




An important property of couplings is closure under sequential
composition.



\begin{lemma}[Sequential composition couplings]\label{lem:sequential-composition-of-couplings}
Let $\mu_1\in\Distr(C_1)$, $\mu_2\in\Distr(C_2)$, $M_1:C_1\rightarrow
\Distr(D_1)$ and $M_2:C_2\rightarrow \Distr(D_2)$. Moreover, let
$R\subseteq C_1\times C_2$ and $S\subseteq D_1\times D_2$. Assume:
$(1)$ $\coupl{\mu_1}{\mu_2}{R}$; and
$(2)$ for every $x_1\in C_1$ and $x_2\in C_2$ such that $R~x_1~x_2$,
  we have $\coupl{M_1(x_1)}{M_2(x_2)}{S}$.
Then $\coupl{(\mathsf{bind}~\mu_1~M_1)}{(\mathsf{bind}~\mu_2~M_2)}{S}$,
where $\mathsf{bind}~\mu~M$ is defined as
$$(\mathsf{bind}~\mu~M)(y) =\sum_x \mu(x) \cdot M(x)(y)$$

\end{lemma}
We conclude this section with the following lemma, which follows from
Strassen's theorem:
\begin{lemma}[Fundamental lemma of couplings]\label{lem:fun-coup}
Let $R\subseteq C_1 \times C_2$, $E_1\subseteq C_1$ and $E_2\subseteq
C_2$ such that for every $x_1\in E_1$ and $x_2\in C_2$, $R~x_1~x_2$
implies $x_2\in E_2$, i.e.\, $R(E_1)\subseteq E_2$. Moreover, let
$\mu_1\in\Distr(C_1)$ and $\mu_2\in\Distr(C_2)$ such that
$\coupl{\mu_1}{\mu_2}{R}$.  Then
$$\Pr_{x_1\leftarrow \mu_1} [x_1\in E_1] \leq \Pr_{x_2\leftarrow\mu_2} [x_2\in E_2]$$
\end{lemma}
This lemma can be used to prove probabilistic inequalities
from the existence of suitable couplings:

\begin{corollary}\label{cor:fundamental}
Let $\mu_1,\mu_2\in\Distr(C)$:
\begin{enumerate}
  \item If $\coupl{\mu_1}{\mu_2}{(=)}$, then for all $x\in C$, $\mu_1(x) = \mu_2(x)$.
\item If $C = \nat$ and $\coupl{\mu_1}{\mu_2}{(\geq)}$, then for all $n\in \nat$,
    $\Pr_{x\leftarrow \mu_1}[x\geq n] \geq \Pr_{x\leftarrow \mu_2}[x\geq n]$
\end{enumerate}
\end{corollary}

In the example at the beginning of the section, the property we want
to prove is precisely that, for every $k$ and $i$, the following
holds:
\[
\Pr_{x_1\leftarrow n_i} [x_1 \geq k] \leq \Pr_{x_2\leftarrow m_i} [x_2 \geq k]
\]
Since we have a $\leq$-coupling, this proof is immediate.
This example is formalized in \autoref{sec:proba-ex}.


\section{Overview of the system}
\label{sec:overview}

In this section we give a high-level overview of our system, with the
details on sections \ref{sec:syntax}, \ref{sec:ghol} and \ref{sec:grhol}.
We start by presenting the base logic, and then we show how to extend it with
probabilities and how to build a relational reasoning system on top of it.

\subsection{Base logic: Guarded Higher-Order Logic}

Our starting point is the Guarded Higher-Order Logic~\cite{CBGB16}
(Guarded HOL)
inspired by the topos of
trees. In addition to the usual constructs of HOL to reason about lambda
terms, this logic features the $\later$ and $\square$ modalities to reason
about infinite terms, in particular streams. The $\later$ modality is used
to reason about objects that will be available in the future, such as tails
of streams. For instance, suppose we want to define an $\All(s,\phi)$ predicate, expressing that
all elements of a stream $s \equiv \cons{n}{xs}$ satisfy a property $\phi$. This can be axiomatized as
follows:
\[\forall (xs: \later \Str{\nat}) (n : \nat). \phi\ n \Rightarrow \later[s\ot
  xs]{\All(s, x. \phi)} \Rightarrow \All(\cons{n}{xs}, x. \phi)\]
We use $x. \phi$ to denote that the formula $\phi$ depends on a free variable
$x$, which will get replaced by the first argument of $\All$.
We have two antecedents. The first one states that the head $n$ satisfies
$\phi$. The second one, $\later[s\ot xs]{\All(s, x. \phi)}$, states that all
elements of $xs$ satisfy $\phi$. Formally, $xs$ is the tail of the stream and will
be available in the future, so it has type $\later\Str{\nat}$.
The \emph{delayed substitution} $\triangleright[s\ot xs]$ replaces $s$ of type
$\Str{\nat}$ with $xs$ of type $\later\Str{\nat}$ inside $\All$ and shifts the
whole formula one step into the future.
In other words, $\later[s\ot xs]{\All(s, x. \phi)}$
states that $\All(-, x.\phi)$ will be satisfied by $xs$ in the future, once
it is available.







\subsection{A system for relational reasoning}
\label{sec:overview:grhol}

When proving relational properties it
is often convenient to build proofs guided by the syntactic structure
of the two expressions to be related. This style of reasoning is
particularly appealing when the two expressions have the same
structure and control-flow, and is appealingly close to the
traditional style of reasoning supported by refinement types. At the
same time, a strict
adherence to the syntax-directed discipline is detrimental to the
expressiveness of the system; for instance, it makes it difficult or
even impossible to reason about structurally dissimilar terms. To
achieve the best of both worlds, we present a relational proof system built
on top of Guarded HOL, which we call Guarded RHOL. Judgements have the shape:
\[\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi}\]
where $\phi$ is a logical formula that may contain two distinguished
variables $\res\ltag$ and $\res\rtag$ that respectively represent the
expressions $t_1$ and $t_2$. This judgement subsumes two typing
judgements on $t_1$ and $t_2$ and a relation $\phi$ on these two
expressions. However, this form of judgement does not tie the logical
property to the type of the expressions, and is key to achieving
flexibility while supporting syntax-directed proofs whenever needed.
The proof system combines rules of two different flavours: two-sided
rules, which relate expressions with the same top-level constructs,
and one-sided rules, which operate on a single expression.

We then extend Guarded HOL with a modality $\diamond$
that lifts assertions over discrete types $C_1$ and $C_2$ to assertions over
$\Distr(C_1)$ and $\Distr(C_2)$.
Concretely, we define for every assertion $\phi$, variables
$x_1$ and $x_2$ of type $C_1$ and $C_2$ respectively, and expressions
$t_1$ and $t_2$ of type $\Distr(C_1)$ and $\Distr (C_2)$ respectively,
the modal assertion $\diamond_{ [x_1\leftarrow t_1,x_2\leftarrow t_2]}
\phi$ which holds iff the interpretations of $t_1$ and $t_2$ are
related by the probabilistic lifting of the interpretation of
$\phi$. We call this new logic Probabilistic Guarded HOL.



We accordingly extend the relational proof system to support reasoning
about probabilistic expressions by adding judgements of the form:
\[\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Distr(C_1)}
          {t_2}{\Distr(C_2)}{
            \diamond_{[x_1\leftarrow \res\ltag, x_2\leftarrow \res\rtag]} \phi}\]
expressing that $t_1$ and $t_2$ are distributions related by a
$\phi$-coupling. We call this proof system Probabilistic Guarded RHOL.
These judgements can be built by using the following rule,
that lifts relational judgements over discrete types
$C_1$ and $C_2$ to judgements over distribution types $\Distr(C_1)$
and $\Distr(C_2)$ when the premises of Strassen's theorem are
satisfied.
\[
\infer[\sf COUPLING]{
       \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Distr(C_1)}{t_2}{\Distr(C_2)}{
         \diamond_{[y_1\leftarrow \res_1, y_2\leftarrow
             \res_2]}\phi}}{
       \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\forall X_1 \subseteq C_1.
        \Pr_{y_1\ot t_1} [y_1\in X_1] \leq \Pr_{y_2\ot t_2} [\exists y_1 \in X_1. \phi]}}
\]



Recall that (discrete time) Markov chains are ``memoryless''
probabilistic processes, whose specification is given by a (discrete)
set $C$ of states, an initial state $s_0$ and a probabilistic
transition function $\mathsf{step}:C \rightarrow \Distr(C)$, where
$\Distr(S)$ represents the set of discrete distributions over $C$.  As
explained in the introduction, a convenient modelling of Markov chains
is by means of probabilistic streams, i.e.\, to model a Markov chain
as an element of $\Distr(\Str{S})$, where $S$ is its underlying state
space. To model Markov chains, we introduce a $\markov$ operator with
type $C \to (C \to \Distr(C)) \to \Distr(\Str{C})$ that, given an
initial state and a transition function, returns a Markov chain. We
can reason about Markov chains by the \rname{Markov} rule (the context, omitted,
does not change):
\begin{gather*}
\inferrule*[right=\sf Markov]
      {\jgrholnoc{t_1}{C_1}{t_2}{C_2}{\phi} \\\\
       \jgrholnoc{h_1}{C_1 \to \Distr(C_1)}{h_2}{C_2 \to \Distr(C_2)}{\psi_3} \\\\
         \vdash \psi_4}
      {\jgrholnoc{\operatorname{markov}(t_1,h_1)}{\Distr(\Str{D_1})}{\operatorname{markov}(t_2,h_2)}{\Distr(\Str{D_2})}
              {\diamond_{\left[\substack{y_1 \ot \res\ltag \\ y_2 \ot \res\rtag}\right]}\phi'}}
\\[0.5em]
\text{ where }
\begin{cases}
  \psi_3 \equiv \forall x_1 x_2. \phi\defsubst{x_1}{x_2} \Rightarrow
     \diamond_{[ y_1 \ot \res\ltag\ x_1, y_2 \ot \res\rtag\ x_2]}\phi\defsubst{y_1}{y_2} \\
  \psi_4 \equiv \forall x_1\ x_2\ xs_1\ xs_2. \phi\defsubst{x_1}{x_2}
      \Rightarrow \later[y_1 \ot xs_1, y_2 \ot xs_2]{\phi'} \Rightarrow \\
      \quad\quad\quad\phi'\subst{y_1}{\cons{x_1}{xs_1}}\subst{y_2}{\cons{x_2}{xs_2}}
\end{cases}
\end{gather*}
Informally, the rule stipulates the existence of an invariant $\phi$
over states. The first premise insists that the invariant hold on the
initial states, the condition $\psi_3$ states that the transition
functions preserve the invariant, and $\psi_4$ states that the
invariant $\phi$ over pairs of states can be lifted to a stream
property $\phi'$.



Other rules of the logic are given in
Figure~\ref{fig:phol}. The language construct $\mathsf{munit}$ creates a point distribution
whose entire mass is at its argument. Accordingly, the [\textsf{UNIT}] rule creates a
straightforward coupling.
The [\textsf{MLET}] rule internalizes sequential composition of
couplings (Lemma~\ref{lem:sequential-composition-of-couplings}) into
the proof system. The construct $\mlet{x}{t}{t'}$ composes a
distribution $t$ with a probabilistic computation $t'$ with one free
variable $x$ by sampling $x$ from $t$ and running $t'$.
The [\textsf{MLET-L}] rule supports one-sided reasoning about
$\mlet{x}{t}{t'}$ and relies on the fact that couplings are closed
under convex combinations. Note that one premise of the rule uses a
unary judgement, with a non-relational modality
$\diamond_{[x\leftarrow \res]} \phi$ whose informal meaning is that
$\phi$ holds with probability $1$ in the distribution $\res$.

The following table summarizes the different base logics we consider,the
relational systems we build on top of them, including the ones presented in~\cite{ABGGS17}, and the equivalences between both sides:
\begin{center}
\small
\begin{tabular}{lcl}
Relational logic  & & Base logic \\ \hline
&&\\[-2mm]
  $\begin{array}{l}\text{RHOL~\cite{ABGGS17}} \\
                   \Gamma \mid \Psi \vdash t_1 \sim t_2 \mid \phi \end{array} $
                    & $\stackrel{\text{\cite{ABGGS17}}}{\Longleftrightarrow}$ 	&
  $\begin{array}{l} \text{HOL~\cite{ABGGS17}} \\
  \Gamma \mid \Psi \vdash \phi\defsubst{t_1}{t_2} \end{array} $ \\[1em]

  $\begin{array}{l} \text{Guarded RHOL~\S\ref{sec:grhol}} \\
                   \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash t_1 \sim t_2 \mid \phi \end{array} $
                    & $\stackrel{\text{Thm~\ref{thm:equiv-rhol-hol}}}{\Longleftrightarrow}$ 	&
  $\begin{array}{l} \text{Guarded HOL~\cite{CBGB16}} \\
    \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash \phi\defsubst{t_1}{t_2} \end{array} $ \\[1em]

  $\begin{array}{l} \text{Probabilistic Guarded RHOL~\S\ref{sec:grhol}} \\
    \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash t_1 \sim t_2 \mid
    \diamond_{[y_1 \ot \res\ltag, y_2\ot\res\rtag]}.\phi \end{array} $
                    & \hspace{.3cm}
                      $\stackrel{\text{Thm~\ref{thm:equiv-rhol-hol}}}{\Longleftrightarrow}$
                       \hspace{.3cm}	&
  $\begin{array}{l} \text{Probabilistic Guarded HOL~\S\ref{sec:ghol}} \\
    \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash
    \diamond_{[y_1 \ot t_1, y_2 \ot t_2]}.\phi \end{array} $
\end{tabular}
\end{center}


\begin{figure*}[!tb]
\small
\begin{gather*}
 \infer[\sf UNIT]{\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\munit{t_1}}{\Distr(C_1)}
          {\munit{t_2}}{\Distr(C_2)}{
            \diamond_{[x_1\leftarrow \res\ltag, x_2\leftarrow \res\rtag]} \phi}} 
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{C_1}{t_2}{C_2}{\phi
          [\res\ltag/x_1, \res\rtag/x_2]}}
\\[0.2em]
\infer[\sf MLET]{
      \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\mlet{x_1}{t_1}{t'_1}}{
      \Distr(D_1)}{\mlet{x_2}{t_2}{t'_2}}{\Distr(D_2)}{\diamond_{\left[\substack{y_1\leftarrow \res_1 \\ y_2\leftarrow \res_2}\right]} \psi}}
    {\begin{array}{c}\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Distr(C_1)}{t_2}{\Distr(C_2)}{
     \diamond_{[x_1\leftarrow \res_1, x_2\leftarrow \res_2]} \phi} \\
     \jgrhol{\Delta}{\Sigma}{\Gamma,x_1:C_1,x_2:C_2}{\Psi,\phi
     }{t'_1}{\Distr(D_1)} {t'_2}{\Distr(D_2)}{\diamond_{[y_1\leftarrow \res_1, y_2\leftarrow \res_2]} \psi}
    \end{array}}
\\[0.2em]
\infer[\sf MLET-L]{
   \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\mlet{x_1}{t_1}{t'_1}}{\Distr(D_1)}{t'_2}{\Distr(D_2)}{
     \diamond_{[y_1\leftarrow \res_1, y_2\leftarrow \res_2]} \psi}}
     {\begin{array}{c}
         \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Distr(C_1)}{\diamond_{[x\leftarrow \res]} \phi}
\\
     \jgrhol{\Delta}{\Sigma}{\Gamma,x_1:C_1}{\Psi,\phi
     }{t'_1}{\Distr(D_1)}{t'_2}{\Distr(D_2)}{\diamond_{[y_1\leftarrow \res_1, y_2\leftarrow \res_2]} \psi}
     \end{array}}
\end{gather*}     
\shrinkcaption
\caption{Proof rules for probabilistic constructs}\label{fig:phol}
\end{figure*}


\subsection{Examples}\label{sec:proba-ex}
We formalize elementary examples from the literature on security and
Markov chains. None of these examples can be verified in prior
systems. Uniformity of \emph{one-time pad} and lumping of \emph{random
  walks} cannot even be stated in prior systems because the two
related expressions in these examples have different types. The
\emph{random walk vs lazy random walk} (shift coupling) cannot be
proved in prior systems because it requires either asynchronous
reasoning or code rewriting.  Finally, the \emph{biased coin example}
(stochastic dominance) cannot be proved in prior work because it
requires Strassen's formulation of the existence of coupling (rather
than a bijection-based formulation) or code rewriting.  We give
additional details below. 

\subsubsection{One-time pad/probabilistic non-interference}
Non-interference~\cite{GoguenM82} is a baseline information flow
policy that is often used to model confidentiality of computations. In
its simplest form, non-interference distinguishes between public (or
low) and private (or high) variables and expressions, and requires
that the result of a public expression not depend on the value of its
private parameters. This definition naturally extends to probabilistic
expressions, except that in this case the evaluation of an expression
yields a distribution rather than a value. There are deep connections
between probabilistic non-interference and several notions of
(information-theoretic) security from cryptography. In this paragraph,
we illustrate different flavours of security properties for one-time
pad encryption. Similar reasoning can be carried out for proving
(passive) security of secure multiparty computation algorithms in the
3-party or multi-party setting~\cite{BogdanovNTW12,Cramer:2015:SMC}.

One-time pad is a perfectly secure symmetric encryption scheme. Its
space of plaintexts, ciphertexts and keys is the set
$\{0,1\}^\ell$---fixed-length bitstrings of size $\ell$. The
encryption algorithm is parametrized by a key $k$---sampled uniformly
over the set of bitstrings $\{ 0,1 \}^\ell$---and maps every plaintext
$m$ to the ciphertext $c = k \oplus m$, where the operator $\oplus$
denotes bitwise exclusive-or on bitstrings. We let $\mathsf{otp}$
denote the expression $\lambda
m. \mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\munit{k\oplus m}}$, where
$\mathcal{U}_{X}$ is the uniform distribution over a finite set $X$.

One-time pad achieves perfect security, i.e.\, the distributions of
ciphertexts is independent of the plaintext. Perfect security can be
captured as a probabilistic non-interference property: 
$$\jgrholnoc{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1
  \}^\ell)}{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)}{
  \forall m_1m_2.
\res\ltag~m_1 \stackrel{\diamond}{=} \res\rtag~m_2
}$$
where $e_1 \stackrel{\diamond}{=} e_2$ is used as a shorthand for
$\diamond_{[y_1\leftarrow e_1, y_2\leftarrow e_2]} y_1 = y_2$. The
crux of the proof is to establish
$$m_1,m_2: \{ 0,1 \}^\ell
\jgrholnoc{\mathcal{U}_{\{0,1\}^\ell}}{\Distr(\{ 0, 1 \}^\ell)}{
  \mathcal{U}_{\{0,1\}^\ell}}{\Distr(\{ 0, 1 \}^\ell)}{
\res\ltag \oplus m_2 \stackrel{\diamond}{=} \res\rtag \oplus m_1
}$$
using the [\textsf{COUPLING}] rule. It suffices to observe that the
assertion induces a bijection, so the image of an arbitrary set $X$
under the relation has the same cardinality as $X$, and hence their
probabilities w.r.t.\, the uniform distributions are equal. One can
then conclude the proof by applying the rules for monadic
sequenciation (\rname{MLET}) and abstraction (rule \rname{ABS} in appendix), using
algebraic properties of $\oplus$.

Interestingly, one can prove a stronger property: rather than proving
that the ciphertext is independent of the plaintext, one can prove
that the distribution of ciphertexts is uniform. This is captured by
the following judgement: 
$$c_1, c_2: \{ 0,1 \}^\ell
\jgrholnoc{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)}
{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)}{\psi}$$
where
$\psi\defeq \forall m_1\,m_2. m_1=m_2\Rightarrow
  \diamond_{[y_1\leftarrow \res\ltag~m_1, y_2\leftarrow \res\rtag ~m_2]} 
  y_1=c_1  \Leftrightarrow y_2=c_2$.
This style of modelling uniformity as a relational property is
inspired from~\cite{BartheEGHS17}. The proof is similar to the
previous one and omitted. However, it is arguably more natural to
model uniformity of the distribution of ciphertexts by the judgement:
$$\jgrholnoc{\mathsf{otp}}{\{ 0,1 \}^\ell \rightarrow \Distr(\{ 0, 1 \}^\ell)}
{\mathcal{U}_{\{0,1\}^\ell}}{\Distr(\{ 0, 1 \}^\ell)}{
  \forall m.~
\res\ltag~m \stackrel{\diamond}{=} \res\rtag
}$$
This judgement is closer to the simulation-based notion of security
that is used pervasively in cryptography, and notably in Universal
Composability~\cite{canetti2001universally}. Specifically, the
statement captures the fact that the one-time pad algorithm can be
simulated without access to the message.  It is interesting to note
that the judgement above (and more generally simulation-based security)
could not be expressed in prior works, since the two expressions of
the judgement have different types---note that in this specific case,
the right expression is a distribution but in the general case the
right expression will also be a function, and its domain will be a
projection of the domain of the left expression.



The proof proceeds as follows. First, we prove
$$
\jgrholnocnot{\mathcal{U}_{\{0,1\}^\ell}}{
\mathcal{U}_{\{0,1\}^\ell}}{
\forall m.~\diamond_{[y_1\leftarrow \res\ltag, y_2\leftarrow \res\rtag]}
  y_1 \oplus m = y_2}$$
using the [\textsf{COUPLING}] rule. Then, we apply the [\textsf{MLET}]
rule to obtain
$$\jgrholnocnot
  {\begin{array}{l}\mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\\ \munit{k\oplus m}} \end{array}}
  {\begin{array}{l}\mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\\ \munit{k}}\end{array}}
     {\diamond_{\left[y_1\leftarrow \res\ltag,
      y_2\leftarrow \res\rtag \right]} y_1 = y_2}$$
We have $\mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\munit{k}} \equiv
\mathcal{U}_{\{0,1\}^\ell}$; hence by equivalence (rule
\rname{Equiv} in appendix), this entails
$$
\jgrholnocnot{\mlet{k}{\mathcal{U}_{\{0,1\}^\ell}}{\munit{k\oplus m}}}
             {\mathcal{U}_{\{0,1\}^\ell}}{\diamond_{[y_1\leftarrow
                   \res\ltag, y_2\leftarrow \res\rtag]} y_1 = y_2}$$
We conclude by applying the one-sided rule for abstraction.



\subsubsection{Stochastic dominance}
Stochastic dominance defines a partial order between random variables
whose underlying set is itself a partial order; it has many different
applications in statistical biology (e.g.\ in the analysis of the
birth-and-death processes), statistical physics (e.g.\ in percolation
theory), and economics. First-order stochastic dominance, which we
define below, is also an important application of probabilistic
couplings. We demonstrate how to use our proof system for proving
(first-order) stochastic dominance for a simple Markov process which
samples biased coins. While the example is elementary, the proof
method extends to more complex examples of stochastic dominance, and
illustrates the benefits of Strassen's formulation of the coupling
rule over alternative formulations stipulating the existence of
bijections (explained later).

We start by recalling the definition of (first-order) stochastic
dominance for the $\mathbb{N}$-valued case. The definition extends to
arbitrary partial orders.
\begin{definition}[Stochastic dominance]
Let $\mu_1,\mu_2\in \Distr(\mathbb{N})$. We say that $\mu_2$
stochastically dominates $\mu_1$, written $\mu_1\leq_{\mathrm{SD}}
\mu_2$, iff for every $n\in\mathbb{N}$,
$$\Pr_{x\leftarrow \mu_1}[x\geq n] \leq \Pr_{x\leftarrow \mu_2}[x\geq n]$$
\end{definition}
The following result, equivalent to \autoref{cor:fundamental},
characterizes stochastic dominance using probabilistic couplings.
\begin{proposition}
Let $\mu_1,\mu_2\in \Distr(\mathbb{N})$. Then $\mu_1\leq_{\mathrm{SD}}
\mu_2$ iff $\coupl{\mu_1}{\mu_2}{(\leq)}$.
\end{proposition}


We now turn to the definition of the Markov chain. For $p\in [0,1]$,
we consider the parametric $\mathbb{N}$-valued Markov chain
$\mathsf{coins} \defeq \markov(0,h)$, with initial state $0$ and
(parametric) step function:
$$ h \defeq \lambda x. \mlet{b}{\bern{p}}{\munit{x+b}}
$$
where, for $p \in [0,1]$, $\bern{p}$ is the Bernoulli distribution on
$\{0,1\}$ with probability $p$ for $1$ and $1-p$ for $0$. Our goal is
to establish that $\mathsf{coins}$ is monotonic, i.e.\, for every
$p_1,p_2\in [0,1]$, $p_1\leq p_2$ implies $\mathsf{coins}~p_1
\leq_{\mathrm{SD}} \mathsf{coins}~p_2$. We formalize this statement as
$$
\jgrholnoc{\mathsf{coins}}{[0,1] \rightarrow \Distr(\Str{\mathbb{N}})}
          {\mathsf{coins}}{[0,1] \rightarrow \Distr(\Str{\mathbb{N}})}
          {\psi}
          $$
where $\psi\defeq \forall p_1,p_2. p_1\leq p_2 \Rightarrow
\diamond_{[y_1 \ot \res\ltag, y_2 \ot \res\rtag]} \All(y_1, y_2, z_1.z_2.z_1\leq z_2)$.
The crux of the proof is to establish stochastic dominance for the
Bernoulli distribution:
$$
p_1:[0,1],p_2:[0,1]\mid p_1\leq p_2 \jgrholnoc{\bern{p_1}}
          {\Distr(\mathbb{N})}
          {\bern{p_2}}
          {\Distr(\mathbb{N})}
          {\res\ltag\stackrel{\diamond}{\leq} \res\rtag}
$$
where we use $e_1 \stackrel{\diamond}{\leq} e_2$ as shorthand for
$\diamond_{[y_1 \ot e_1, y_2 \ot e_2]} y_1\leq y_2$. This is proved
directly by the \rname{COUPLING} rule and checking by simple
calculations that the premise of the rule is valid.

We briefly explain how to conclude the proof. Let $h_1$ and $h_2$ be
the step functions for $p_1$ and $p_2$ respectively. It is clear from
the above that (context omitted):
$$
x_1\leq x_2 \jgrholnoc{h_1\ x_1}{\Distr(\bool)}
         {h_2\ x_2}{\Distr(\bool)}
         {\diamond_{[y_1 \ot \res\ltag, y_2 \ot \res\rtag]}.
           {y_1\leq y_2}}
$$
and by the definition of $\All$:
$$x_1 \leq x_2 \Rightarrow {\All(xs_1,xs_2,z_1.z_2.z_1\leq z_2)}
\Rightarrow \All(\cons{x_1}{\later xs_1}, \cons{x_2}{\later xs_2}, z_1.z_2.z_1\leq z_2)$$
So, we can conclude by applying the \rname{Markov} rule.

It is instructive to compare our proof with prior formalizations, and
in particular with the proof in \cite{BartheEGHSS15}. Their proof is
carried out in the \textsf{pRHL} logic, whose [\textsf{COUPLING}] rule
is based on the existence of a bijection that satisfies some property,
rather than on our formalization based on Strassen's Theorem. Their
rule is motivated by applications in cryptography, and works well for
many examples, but is inconvenient for our example at hand, which
involves non-uniform probabilities. Indeed, their proof is based on
code rewriting, and is done in two steps. First, they prove equivalence between sampling and
returning $x_1$ from $\bern{p_1}$; and sampling $z_1$ from
$\bern{p_2}$, $z_2$ from $\bern{\rfrac{p_1}{p_2}}$ and returning
$z= z_1 \land z_2$. Then, they find a coupling between $z$ and
$\bern{p_2}$.


\subsubsection{Shift coupling: random walk vs lazy random walk}
The previous example is an instance of a lockstep coupling, in that it
relates the $k$-th element of the first chain with the $k$-th element
of the second chain. Many examples from the literature follow this
lockstep pattern; however, it is not always possible to establish
lockstep couplings. Shift couplings are a relaxation of lockstep
couplings where we relate elements of the first and second chains
without the requirement that their positions coincide.

We consider a simple example that motivates the use of shift
couplings.  Consider the random walk and lazy random walk (which, at
each time step, either chooses to move or stay put), both defined as
Markov chains over $\mathbb{Z}$. For simplicity, assume that both
walks start at position 0. It is not immediate to find a coupling
between the two walks, since the two walks necessarily get
desynchronized whenever the lazy walk stays put. Instead, the trick is
to consider a lazy random walk that moves two steps instead of
one. The random walk and the lazy random walk of step 2 are defined by
the step functions:
$$\begin{array}{rcl}
  \operatorname{step} & \defeq & \lambda x.\mlet{z}{\mathcal{U}_{\{-1,1\}}}{\munit{z+x}} \\
  \operatorname{lstep2} & \defeq & \lambda x.\mlet{z}{\mathcal{U}_{\{-1,1\}}}{\mlet{b}{\mathcal{U}_{\{0,1\}}}{\munit{x+2*z*b}}}
\end{array}$$
After 2 iterations of $\operatorname{step}$, the position has either
changed two steps to the left or to the right, or has returned to the
initial position, which is the same behaviour $\operatorname{lstep2}$
has on every iteration.  Therefore, the coupling we want to find should
equate the elements at position $2i$ in $\operatorname{step}$ with the
elements at position $i$ in $\operatorname{lstep2}$. The details on how
to prove the existence of this coupling are in \autoref{sec:grhol}.

\subsubsection{Lumped coupling: random walks on 3 and 4 dimensions}
A Markov chain is \emph{recurrent} if it has probability 1 of
returning to its initial state, and \emph{transient} otherwise. It is
relatively easy to show that the random walk over $\mathbb{Z}$ is
recurrent. One can also show that the random walk over $\mathbb{Z}^2$
is recurrent.  However, the random walk over $\mathbb{Z}^3$ is
transient.


For higher dimensions, we can use a coupling
argument to prove transience. Specifically, we can define a coupling
between a lazy random walk in $n$ dimensions and a random walk in $n
+m$ dimensions, and derive transience of the latter from transience of
the former. We define the (lazy) random walks below, and sketch the
coupling arguments.

Specifically, we show here the particular case of the transience of
the 4-dimensional random walk from the transience of the 3-dimensional
lazy random walk.  We start by defining the stepping functions:
\[\begin{array}{rl}
  \operatorname{step}_4 &: \mathbb{Z}^4 \to \Distr(\mathbb{Z}^4)
  \defeq \lambda z_1. \mlet{x_1}{\unif{U_4}}{\munit{z_1 +_4 x_1}}
  \\
  \operatorname{lstep}_3 &: \mathbb{Z}^3 \to \Distr(\mathbb{Z}^3)
  \defeq \lambda z_2. \mlet{x_2}{\unif{U_3}}{\mlet{b_2}{\bern{\rfrac{3}{4}}}{\munit{z_2 +_3 b_2*x_2 }}}
  \end{array}
\]
where $U_i=\{(\pm 1,0,\dots 0), \dots, (0,\dots,0,\pm 1)\}$ are the
vectors of the basis of $\mathbb{Z}^i$ and their opposites.  Then, the
random walk of dimension 4 is modelled by $\operatorname{rwalk4} \defeq
\markov(0, \operatorname{step_4})$, and the lazy walk of dimension 3 is
modelled by $\operatorname{lwalk3} \defeq
\markov(0, \operatorname{step_3})$.  We want to prove:
$$
\jgrholnoc{\operatorname{rwalk4}}{\Distr(\Str{\mathbb{Z}^{4}})}
       {\operatorname{lwalk3}}{\Distr(\Str{\mathbb{Z}^{3}})}
       {\diamond_{\left[\substack{y_1\leftarrow \res\ltag \\ y_2\leftarrow \res\rtag}\right]} 
  \All(y_1, y_2, z_1.z_2.\operatorname{pr}^{4}_{3}(z_1) = z_2)}
$$
where $\operatorname{pr}^{n_2}_{n_1}$ denotes the standard projection from
$\mathbb{Z}^{n_2}$ to $\mathbb{Z}^{n_1}$. 

We apply the \rname{Markov} rule. The only interesting premise requires proving that the transition function preserves the coupling:
\[p_2=\operatorname{pr}^{4}_{3}(p_1) \vdash \operatorname{step_4} \sim \operatorname{lstep}_3
                    \mid \forall x_1 x_2. x_2=\operatorname{pr}^4_3(x_1) \Rightarrow \diamond_{\left[\substack{y_1\ot \res\ltag\ x_1 \\ y_2\ot \res\rtag\ x_2}\right]} \operatorname{pr}^4_3(y_1)=y_2 \]




To prove this, we need to find the appropriate coupling, i.e., one
that preserves the equality.  The idea is that the step in
$\mathbb{Z}^3$ must be the projection of the step in
$\mathbb{Z}^4$. This corresponds to the following judgement:
\[\left.\begin{array}{rl}
   \lambda z_1. &\mlet{x_1}{\unif{U_4}}
               {\\ &\munit{z_1 +_4 x_1}}
  \end{array}
  \sim
  \begin{array}{rl}
  \lambda z_2. &\mlet{x_2}{\unif{U_3}}
               {\\ &\mlet{b_2}{\bern{\rfrac{3}{4}}}
               {\\ &\munit{z_2 +_3 b_2*x_2}}}
  \end{array}
  \; \right| \;
  \begin{array}{c}
  \forall z_1 z_2. \operatorname{pr}^4_3(z_1) = z_2 \Rightarrow \\ \operatorname{pr}^4_3(\res\ltag\ z_1) \stackrel{\diamond}{=} \res\rtag\ z_2
  \end{array}\]
which by simple equational reasoning is the same as
\[\left.\begin{array}{rl}
   \lambda z_1. &\mlet{x_1}{\unif{U_4}}
               {\\ &\munit{z_1 +_4 x_1}}
  \end{array}
  \sim
  \begin{array}{rl}
  \lambda z_2. &\mlet{p_2}{\unif{U_3} \times \bern{\rfrac{3}{4}}}
               {\\ &\munit{z_2 +_3 \pi_1(p_2)*\pi_2(p_2)}}
  \end{array}
  \; \right| \;
  \begin{array}{c}
  \forall z_1 z_2. \operatorname{pr}^4_3(z_1) = z_2 \Rightarrow \\ \operatorname{pr}^4_3(\res\ltag\ z_1) \stackrel{\diamond}{=} \res\rtag\ z_2
  \end{array}
  \]

We want to build a coupling such that if we sample $(0,0,0,1)$ or
$(0,0,0,-1)$ from $\unif{U_3}$, then we sample $0$ from
$\bern{\rfrac{3}{4}}$, and otherwise if we sample
$(x_1,x_2,x_3,0)$ from $\unif{U_4}$, we sample $(x_1,x_2,x_3)$ from
$U_3$. Formally, we prove this with the \rname{Coupling} rule.
Given $X:U_4 \to \bool$, by simple computation we show that:
\[ \Pr_{z_1 \sim \unif{U_4}}[z_1 \in X] \leq \Pr_{z_2 \sim \unif{U_3} \times \bern{\rfrac{3}{4}}}[z_2 \in \{ y \mid \exists x \in X. {\sf pr}^4_3(x) = \pi_1(y)*\pi_2(y) \}]\]



This concludes the proof. From the previous example, it follows that
the lazy walk in 3 dimensions is transient, since the random walk in
3 dimensions is transient. By simple reasoning, we now conclude that
the random walk in 4 dimensions is also transient.





 
\section{Probabilistic Guarded Lambda Calculus}
\label{sec:syntax}





To ensure that a function on infinite datatypes is well-defined, one
must check that it is \emph{productive}. This means that any finite
prefix of the output can be computed in finite time. For instance,
consider the following function on streams:
\[
 \mathtt{letrec\ bad\ (x:xs) = x : tail (bad\ xs)}
\]
This function is not productive since only the first element can be computed. We
can argue this as follows:
Suppose that the tail of a stream is available one unit of time after its head,
and that that \texttt{x:xs} is available at time 0. How much time does it take
for \texttt{bad} to start outputting its tail? Assume it takes $k$ units of
time. This means that \texttt{tail(bad\ xs)} will be available at time $k+1$ ,
since \texttt{xs} is only available at time 1. But \texttt{tail(bad\ xs)} is
exactly the tail of \texttt{bad(x:xs)}, and this is a contradiction, since
\texttt{x:xs} is available at time 0 and therefore the tail of
\texttt{bad(x:xs)} should be available at time $k$. Therefore, the tail of
\texttt{bad} will never be available.

The guarded lambda calculus solves the productivity problem by
distinguishing at type level between data that is available now and
data that will be available in the future, and restricting when
fixpoints can be defined.  Specifically, the guarded lambda calculus
extends the usual simply typed lambda calculus with two modalities:
$\later$ (pronounced \textit{later}) and $\square$
(\textit{constant}).  The later modality represents data that will be
available one step in the future, and is introduced and removed by the
term formers $\later$ and $\prev$\! respectively.  This modality is
used to guard recursive occurrences, so for the calculus to remain
productive, we must restrict when it can be eliminated.  This is
achieved via the constant modality, which expresses that all the data
is available at all times.  In the remainder of this section we present a probabilistic extension
of this calculus.


\paragraph*{Syntax}
Types of the calculus are defined by the grammar
\begin{align*}
  A,B ::= b \mid \nat \mid A \times B \mid A + B  \mid A \to B \mid \Str{A} \mid \square~A \mid \later{A} \mid \Distr(C)
\end{align*}
where $b$ ranges over a collection of base types.
$\Str{A}$ is the type of guarded streams of elements of type
$A$. Formally, the type $\Str{A}$ is isomorphic to $A \times
\later{\Str{A}}$. This isomorphism gives a way to introduce streams
with the function $(\cons{}{}) : A \to \later\Str{A} \to
\Str{A}$ and to eliminate them with the functions $\hd: \Str{A} \to A$
and $\tl: \Str{A} \to \later\Str{A}$.
$\Distr (C)$ is the type of distributions over \emph{discrete types}
$C$.  Discrete types are defined by the following grammar, where $b_0$
are discrete base types, e.g., $\zint$.
\begin{align*}
  C,D ::= b_0 \mid \nat \mid C \times D \mid C + D \mid  \Str{C} | \later{C}.
\end{align*}
Note that, in particular, arrow types are not discrete but streams
are. This is due to the semantics of streams as sets of finite
approximations, which we describe in the next subsection. Also note
that $\square\Str{A}$ is not discrete since it makes the full infinite
streams available.

We also need to distinguish between arbitrary types $A, B$ and
constant types $S, T$, which are defined by the following grammar
\begin{align*}
  S, T ::= b_C \mid \nat \mid S \times T \mid S + T  \mid S \to T \mid \square~A
\end{align*}
where $b_C$ is a collection of constant base types.  Note in
particular that for any type $A$ the type $\square~A$ is constant.

The terms of the language $t$ are defined by the following grammar
\begin{align*}
  t &::= ~x
        \mid c
        \mid 0 \mid S t \mid \casenat{t}{t}{t} \mid \mu \mid \munit{t} \mid \mlet{x}{t}{t} \\
&\mid \langle t, t \rangle \mid \pi_1 t \mid \pi_2 t
        \mid \inj{1} t \mid \inj{2} t \mid \CASE t \OF \inj{1} x . t; \inj{2} y . t
        \mid \lambda x . t \mid t\,t \mid \fix{x}{t}\\
        &\mid \cons{t}{ts} \mid \hd t \mid \tl t
        \mid \boxx{t} \mid \letbox{x}{t}{t} \mid \letconst{x}{t}{t}
        \mid \latern[\xi]t \mid \prev t
\end{align*}
where $\xi$ is a delayed substitution, a sequence of bindings
$\hrt{x_1 \gets t_1, \ldots, x_n \gets t_n}$.
The terms $c$ are constants corresponding to the base types used and
$\munit{t}$ and $\mlet{x}{t}{t}$ are the introduction and sequencing
construct for probability distributions.  The meta-variable $\mu$
stands for base distributions like $\unif{C}$ and $\bern{p}$.

Delayed substitutions were introduced in \cite{BGCMB16} in a dependent type theory to be able to work with types dependent on terms of type $\later{A}$.
In the setting of a simple type theory, such as the one considered in this paper, delayed substitutions are equivalent to having the applicative structure~\cite{McBride:Applicative} $\app$ for the $\later$ modality.
However, delayed substitutions extend uniformly to the level of propositions, and thus we choose to use them in this paper in place of the applicative structure.

\paragraph*{Denotational semantics}

The meaning of terms is given by a denotational model in the category $\trees$ of presheaves over $\omega$, the first infinite ordinal.
This category $\trees$ is also known as the \emph{topos of trees}~\cite{Birkedal-et-al:topos-of-trees}.
In previous work~\cite{CBGB16}, it was shown how to model most of the constructions of the guarded lambda calculus and its internal logic, with the notable exception of the probabilistic features.
Below we give an elementary
presentation of the semantics.

Informally, the idea behind the topos of trees is to represent
(infinite) objects from their finite approximations, which we observe
incrementally as time passes.  Given an object $x$, we can consider a
sequence $\{x_i\}$ of its finite approximations observable at time
$i$. These are trivial for finite objects, such as a natural number,
since for any number $n$, $n_i = n$ at every $i$.  But for infinite
objects such as streams, the $i$th approximation is the prefix of
length $i+1$.

Concretely, the category $\trees$ consists of:
\begin{itemize}
  \item Objects $X$: families of sets $\{X_i\}_{i\in\nat}$ together with \emph{restriction functions} $r_n^X : X_{n+1} \to X_n$.
  We will write simply $r_n$ if $X$ is clear from the context.
  \item Morphisms $X \to Y$ : families of functions $\alpha_n : X_n \to Y_n$ commuting with restriction functions in the sense of $r_n^Y \circ \alpha_{n+1} = \alpha_n \circ r_n^X$.
\end{itemize}


The full interpretation of types of the calculus can be found in
\autoref{fig:sem-types} in the appendix. The main points we want to
highlight are:
\begin{itemize}
  \item Streams over a type $A$ are interpreted as sequences of finite
    prefixes of elements of $A$ with the restriction functions of $A$:
    $$\sem{\Str{A}} \defeq \sem{A}_0 \times \{*\} \xleftarrow{r_0
      \times !} \sem{A}_1 \times \sem{\Str{A}}_0
    \xleftarrow{r_1 \times r_0 \times !}
    \sem{A}_2\times \sem{\Str{A}}_1 \leftarrow \cdots$$

  \item Distributions over a discrete object $C$ are
    defined as a sequence of distributions over each
    $\sem{C}_i$: $$\sem{\Distr(C)} \defeq \Distr(\sem{C}_0)
    \stackrel{\Distr(r_0)}{\longleftarrow} \Distr(\sem{C}_1)
    \stackrel{\Distr(r_1)}{\longleftarrow} \Distr(\sem{C}_2)
    \stackrel{\Distr(r_2)}{\longleftarrow} \ldots,$$ where $\Distr(\sem{C}_i)$ is
    the set of (probability density) functions $\mu : \sem{C}_i \to [0,1]$ such that $\sum_{x_\in X} \mu
    x = 1$, and $\Distr(r_i)$ adds the probability density of all the points
    in $\sem{C}_{i+1}$ that are sent by $r_i$ to the same point in the $\sem{C}_{i}$. In other words,
    $\Distr(r_i)(\mu)(x) = \Pr_{y \ot \mu}[r_i(y) = x]$
\end{itemize}







An important property of the interpretation is that discrete types are interpreted as objects $X$ such that $X_i$ is finite or countably infinite for every $i$.
This allows us to define distributions on these objects without the need for measure theory.
In particular, the type of guarded streams $\Str{A}$ is discrete provided $A$ is, which is clear from the interpretation of the type $\Str{A}$.
Conceptually this holds because $\sem{\Str{A}}_i$ is an approximation of real streams, consisting of only the first $i+1$ elements.

An object $X$ of $\trees$ is \emph{constant} if all its restriction
functions are bijections.  Constant types are interpreted as constant
objects of $\trees$ and for a constant type $A$
the objects $\sem{\square A}$ and $\sem{A}$ are isomorphic in $\trees$.

\paragraph*{Typing rules}

Terms are typed under a dual context $\Delta \mid \Gamma$, where $\Gamma$ is a usual context that binds variables to a type, and $\Delta$ is a constant context containing variables bound to types that are \textit{constant}.
The term $\letconst{x}{u}{t}$ allows us to shift variables between constant and non-constant contexts. The typing rules can be found in \autoref{fig:glambda}.

The semantics of such a dual context $\Delta \mid \Gamma$ is given as the product of types in $\Delta$ and $\Gamma$, except that we implicitly add $\square$ in front of every type in $\Delta$.
In the particular case when both contexts are empty, the semantics of the dual context correspond to the terminal object $1$, which is the singleton set $\{\ast\}$ at each time. 



The interpretation of the well-typed term $\Delta \mid \Gamma \vdash t : A$ is
defined by induction on the typing derivation, and can be found in
\autoref{fig:sem-glc} in the appendix.




\begin{figure*}[!tb]
\small
\begin{gather*}
  \inferrule
  {x : A \in \Gamma}
  {\Delta \mid \Gamma \vdash x : A}
  \qquad
  \inferrule
  {x : A \in \Delta}
  {\Delta \mid \Gamma \vdash x : A}
  \qquad
  \inferrule
  {\Delta \mid \Gamma, x : A \vdash t : B}
  {\Delta \mid \Gamma \vdash \lambda x . t : A \to B}
  \\
  \inferrule
  {\Delta \mid \Gamma \vdash t : A \to B \and \Delta \mid \Gamma \vdash u : A}
  {\Delta \mid \Gamma \vdash t\,u :  B}
  \qquad
  \inferrule
  {\Delta \mid \Gamma, f : \later{A} \vdash t :  A}
  {\Delta \mid \Gamma \vdash \fix{f}{t} : A}
  \qquad
  \inferrule
  {\Delta \mid \cdot \vdash t : \later A}
  {\Delta \mid \Gamma \vdash \prev{t} : A}
  \\
  \inferrule
  {\Delta \mid \cdot \vdash t : A}
  {\Delta \mid \Gamma \vdash \boxx{t} : \square A}
  \qquad
  \inferrule
  {\Delta \mid \Gamma \vdash u : \square B \and
    \Delta, x : B \mid \Gamma \vdash t : A}
  {\Delta \mid \Gamma \vdash \letbox{x}{u}{t} : A}
  \\
  \inferrule
  {\Delta \mid \Gamma \vdash u : B \and 
    \Delta, x : B \mid \Gamma \vdash t : A \and 
    B \text{ constant}}
  {\Delta \mid \Gamma \vdash \letconst{x}{u}{t} : A}
  \\
  \inferrule
  {\Delta \mid \Gamma, x_1 : A_1, \cdots x_n : A_n \vdash t : A
    \and
    \Delta \mid \Gamma \vdash t_i : \later{A_i}}
  {\Delta \mid \Gamma \vdash \later[x_1 \gets t_1, \ldots, x_n \gets t_n]{t} : \later{A}}
  \qquad
  \inferrule
  {\Delta \mid \Gamma \vdash t : A \and A \text{ discrete }}
  {\Delta \mid \Gamma \vdash \munit{t} : \Distr(A)}
  \\
  \inferrule
  {\Delta \mid \Gamma \vdash t : \Distr(A) \and \Delta \mid \Gamma, x : A \vdash u : \Distr(B)}
  {\Delta \mid \Gamma \vdash \mlet{x}{t}{u} : \Distr(B)}
  \qquad
  \inferrule
  {\mu \text{ primitive distribution on type } A}
  {\Delta \mid \Gamma \vdash \mu : \Distr(A)}
\end{gather*}
\shrinkcaption
\caption{A selection of the typing rules of the guarded lambda calculus.
  The rules for products, sums, and natural numbers are standard.}
\label{fig:glambda}
\end{figure*}




\paragraph*{Applicative structure of the later modality}
As in previous work we can define the operator $\app$ satisfying the typing rule
\[
  \inferrule
  {\Delta \mid \Gamma \vdash t : \later{(A \to B)} \and \Delta \mid \Gamma \vdash u : \later{A}}
  {\Delta \mid \Gamma \vdash t \app u : \later{B}}
\]
and the equation
  $(\later{t}) \app (\later{u}) \equiv \later(t\ u)$
as the term
  $t \app u \defeq \latern[\hrt{f \gets t, x \gets u}]{f\,x}$.






\paragraph*{Example: Modelling Markov chains}








As an application of $\app$ and an example of how to use guardedness
and probabilities together, we now give the precise
definition of the $\markov$ construct that we used to model Markov chains earlier:
\[
\begin{array}{rl}
  \markov &: C  \to (C \to \Distr(C)) \to \Distr(\Str{C}) \\
  \markov &\defeq \fix{f}{\lambda x. \lambda h. \\
          &\quad \mlet {z}{h\ x}{ \mlet{t}{\operatorname{swap_{\later\Distr}^{\Str{C}}}(f \app  \later z \app \later h)}
            { \munit{\cons{x}{t}}}}}
\end{array}
\]
The guardedness condition gives $f$ the type $\later( C \to (C \to
\Distr(C)) \to \Distr(\Str{C}))$ in the body of the
fixpoint. Therefore, it needs to be applied functorially (via $\app$)
to $\later z$ and $\later h$, which gives us a term of type $\later
\Distr(\Str{C})$. To complete the definition we need to build a term
of type $\Distr(\later\Str{C})$ and then sequence it with $\cons{}{}$
to build a term of type $\Distr(\Str{C})$. To achieve this, we use the
primitive operator $\operatorname{swap_{\later\Distr}^C} :
\later\Distr(C) \to \Distr(\later C)$, which witnesses the isomorphism
between $\later\Distr(C)$ and $\Distr(\later C)$. For this isomorphism
to exist, it is crucial that distributions be total (i.e., we cannot
use subdistributions). Indeed, the denotation for $\later\Distr(C)$ is
the sequence $\{\ast\} \ot \Distr(C_1) \ot \Distr(C_2) \ot \dots$,
while the denotation for $\Distr(\later C)$ is the sequence
$\Distr(\{\ast\}) \ot \Distr(C_1) \ot \Distr(C_2) \ot \dots$, and
$\{\ast\}$ is isomorphic to $\Distr(\{\ast\})$ in ${\sf Set}$ only if
$\Distr$ considers only total distributions.









 
\section{Guarded higher-order logic}


\label{sec:ghol}
We now introduce Guarded HOL (GHOL), which is a higher-order logic to reason about terms of the guarded lambda calculus.
The logic is essentially that of~\cite{CBGB16}, but presented with the dual context formulation analogous to the dual-context typing judgement of the guarded lambda calculus.
Compared to standard intuitionistic higher-order logic, the logic GHOL has two additional constructs, corresponding to additional constructs in the guarded lambda calculus.
These are the later modality ($\later$) \emph{on propositions}, with delayed substitutions,
which expresses that a proposition holds one time unit into the future, and the ``always''
modality $\square$, which expresses that a proposition holds at all times.
Formulas are defined by the grammar:
\[ \phi, \psi ::= \top \mid \phi \wedge \psi \mid \phi \vee \psi \mid \neg \psi
  \mid \forall x. \phi \mid \exists x. \phi \mid \later[x_1 \ot t_1 \dots x_n
  \ot t_n]\phi \mid \square \phi\]
The basic judgement of the logic is $\Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash \phi$ where $\Sigma$ is a logical context for $\Delta$ (that is, a list of formulas well-formed in $\Delta$) and $\Psi$ is another logical context for the dual context $\Delta\mid\Gamma$.
The formulas in context $\Sigma$ must be \emph{constant} propositions.
We say that a proposition $\phi$ is \emph{constant} if it is well-typed in context $\Delta \mid \cdot$ and moreover if every occurrence of the later modality in $\phi$ is under the $\square$ modality.
Selected rules are displayed in Figure~\ref{fig:hol} on page~\pageref{fig:hol}. We highlight \rname{Loeb} induction, which is the key to reasoning about fixpoints: to prove that $\phi$ holds now, one
can assume that it holds in the future. 
The interpretation of the formula $\Delta \mid \Gamma \vdash \phi$ is a subobject of the interpretation $\sem{\Delta \mid \Gamma}$.
Concretely the interpretation $A$ of $\Delta \mid \Gamma \vdash \phi$ is a family $\left\{A_i\right\}_{i=0}^\infty$ of sets such that $A_i \subseteq \sem{\Delta \mid \Gamma}_i$.
This family must satisfy the property that if $x \in A_{i+1}$ then $r_i(x) \in A_i$ where $r_i$ are the restriction functions of $\sem{\Delta \mid \Gamma}$.
The interpretation of formulas is defined by induction on the typing derivation.
In the interpretation of the context $\Delta \mid \Sigma \mid \Gamma \mid \Psi$ the formulas in $\Sigma$ are interpreted with the added $\square$ modality.
Moreover all formulas $\phi$ in $\Sigma$ are typeable in the context $\Delta \mid \cdot \vdash \phi$ and thus their interpretations are subsets of $\sem{\square \Delta}$.
We treat these subsets of $\sem{\Delta \mid \Gamma}$ in the obvious way.

The cases for the semantics of the judgement $\Delta \mid \Gamma \vdash \phi$ can be found in the appendix.
It can be shown that this logic is sound with respect to its model in the topos of trees.

\begin{theorem}[Soundness of the semantics]\label{thm:sound-ghol}
  The semantics of guarded higher-order logic is sound: if $\Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash \phi$ is derivable then for all $n \in \nat$, $\sem{\square\Sigma}_n \cap \sem{\Psi}_n \subseteq \sem{\phi}$.
\end{theorem}


In addition, Guarded HOL is expressive enough to axiomatize standard probabilities over discrete sets. 
This axiomatization can be used to define the $\diamond$ modality directly in Guarded HOL (as opposed to our relational proof system,
were we use it as a primitive).
Furthermore, we can derive from this axiomatization additional rules to reason about couplings, which can be seen
in Figure~\ref{fig:prob-hol}. These rules will be the key to proving the soundness of the probabilistic fragment of the relational proof system, and can
be shown to be sound themselves.

\begin{proposition}[Soundness of derived rules]\label{thm:sound-prob-ghol}
  The additional rules are sound.
\end{proposition}


\begin{figure*}[!tb]
\small
\begin{gather*}
\infer[\sf AX_U]
       {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi}}
       {\phi \in \Psi}
\quad
\infer[\sf AX_G]
       {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi}}
       {\phi \in \Sigma}
\quad
\infer[\sf CONV]
      {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{t=t'}}
      {\Gamma \vdash t:\tau & \Gamma \vdash t':\tau & t \equiv t'}
\\[0.3em]
\infer[\sf SUBST]
       {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{x}{u}}}
       {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{x}{t}} & \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{t = u}}
\quad\quad
\infer[\sf Loeb]
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi}} 
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi, \later\phi}{\phi}}
\\[0.3em]
\infer[\sf \later_I]
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot t_1,\dots,x_n \ot t_n]\phi}} 
        {\jghol{\Delta}{\Sigma}{\Gamma, x_1:A_1, \dots, x_n:A_n}{\Psi}{\phi} &
          \Delta\mid\Gamma\vdash t_1 : \later A_1 & \dots &
          \Delta\mid\Gamma\vdash t_n : \later A_n}
\\[0.3em]
\infer[\sf \later_E]
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{x_1}{\prev t_1}\dots\subst{x_n}{\prev t_n}}}
        {\jghol{\Delta}{\Sigma}{\cdot}{\cdot}{\later[x_1 \ot t_1 \dots x_n \ot t_n]\phi} &
         \Delta \mid \bullet \vdash t_1 : \later A_1 & \dots & \Delta \mid
         \bullet \vdash t_n : \later A_n}
\\[0.3em]
\inferrule
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1\!\ot\!t_1,\dots,x_n\!\ot\!t_n]\psi} \\
         \Delta \mid \Gamma \vdash t_1 : \later A_1 \; \dots\; \Delta \mid
         \Gamma \vdash t_n : \later A_n \\\\
         \jghol{\Delta}{\Sigma}{\Gamma, x_1:A_1, \dots, x_n:A_n}{\Psi, \psi}{\phi}}
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot t_1,\dots,x_n \ot t_n]\phi}}
 {\sf \later_{App}}
\\[0.3em]
\infer[\sf \square_I]
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square\phi}}
        {\jghol{\Delta}{\Sigma}{\cdot}{\cdot}{\phi}}
\quad\quad
\infer[\sf \square_E]
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi}}  
        {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square\psi} &
         \jghol{\Delta}{\Sigma,\psi}{\Gamma}{\Psi}{\phi}}
\end{gather*}
\shrinkcaption
\caption{Selected Guarded Higher-Order Logic rules}\label{fig:hol}
\end{figure*}

\begin{figure*}[!tb]
\small
\begin{gather*}
\inferrule*[right = \sf MONO2]
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1, x_2 \leftarrow t_2]}\phi} \\
    \jghol{\Delta}{\Sigma}{\Gamma, x_1 : C_1, x_2 : C_2}{\Psi, \phi}{\psi}}
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1, x_2\leftarrow t_2]}\psi}}
\\[0.3em]
\inferrule*[right = \sf UNIT2]
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{x_1}{t_1}\subst{x_2}{t_2}}}
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}
  {\diamond_{[x_1\leftarrow \munit{t_1}, x_2\leftarrow \munit{t_2}]}\phi}}
\\[0.3em]
\inferrule*[right = \sf MLET2]
          {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1, x_2\leftarrow t_2]} \phi}
          \\ \jghol{\Delta}{\Sigma}{\Gamma,x_1:C_1,x_2:C_2}{\Psi,\phi}
            {\diamond_{[y_1\leftarrow t'_1, y_2\leftarrow t'_2]}\psi}}
      {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}
             {\diamond_{[y_1\leftarrow \mlet{x_1}{t_1}{t'_1}, y_2\leftarrow \mlet{x_2}{t_2}{t'_2}]} \psi}}
\\[0.3em]
\inferrule*[right = \sf MLET{-}L]
          {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1]}\phi}
          \\ \jghol{\Delta}{\Sigma}{\Gamma,x_1:C_1}{\Psi,\phi}
            {\diamond_{[y_1\leftarrow t'_1, y_2\leftarrow t'_2]}\psi}}
      {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}
          {\diamond_{[y_1\leftarrow \mlet{x_1}{t_1}{t'_1}, y_2\leftarrow t'_2]} \psi}}
\end{gather*}
\shrinkcaption
\caption{Derived rules for probabilistic constructs}\label{fig:prob-hol}
\end{figure*}


 
\section{Relational proof system}
\label{sec:grhol}

We complete the formal description of the system by describing the
proof rules for the non-probabilistic fragment of the relational proof
system (the rules of the probabilistic fragment were described in
Section~\ref{sec:overview:grhol}).

\subsection{Proof rules}






The rules for core $\lambda$-calculus constructs are
identical to those of~\cite{ABGGS17}; for convenience, we present a selection
of the main rules in Figure \ref{fig:sel-rhol} in the appendix.



We briefly comment on the two-sided rules for the new constructs
(Figure~\ref{fig:s-rhol}).
The notation $\Omega$ abbreviates a context $\Delta\mid\Sigma\mid\Gamma\mid\Psi$.
The rule \rname{Next} relates two terms that have a $\later$ term
constructor at the top level. We require that both have one term in
the delayed substitutions and that they are related pairwise.  Then
this relation is used to prove another relation between the main
terms. This rule can be generalized to terms with more than one term
in the delayed substitution.
The rule \rname{Prev} proves a relation between terms from the same delayed relation by
applying $\mathrm{prev}$ to both terms.
The rule \rname{Box} proves a relation between two boxed terms if the
same relation can be proven in a constant context.
Dually, \rname{LetBox} uses a relation between two boxed terms to prove
a relation between their unboxings.
\rname{LetConst} is similar to \rname{LetBox}, but it requires instead a relation
between two constant terms, rather than explicitly $\square$-ed terms.
The rule \rname{Fix} relates two fixpoints following the \rname{Loeb} rule from Guarded HOL.
Notice that in the premise, the fixpoints need to appear in the delayed substitution so that
the inductive hypothesis is well-formed.
The rule \rname{Cons} proves relations on streams
from relations between their heads and tails, while
\rname{Head} and \rname{Tail} behave as converses of \rname{Cons}.


Figure~\ref{fig:a-rhol} contains the one-sided versions of the rules.
We only present the left-sided versions as the right-sided versions
are completely symmetric. The rule \rname{Next-L} relates at $\phi$ a
term that has a $\later$ with a term that does not have a
$\later$. First, a unary property $\phi'$ is proven on the term $u$ in
the delayed substitution, and it is then used as a premise to prove
$\phi$ on the terms with delays removed. Rules for proving unary
judgements can be found in the appendix.
Similarly, \rname{LetBox-L} proves a unary property on the term that gets unboxed
and then uses it as a precondition.
The rule \rname{Fix-L} builds a fixpoint just on the left, and relates it with
an arbitrary term $t_2$ at a property $\phi$. Since $\phi$ may contain the
variable $\res\rtag$ which is not in the context, it has to be replaced when
adding $\later\phi$ to the logical context in the premise of the rule.
The remaining rules are similar to their two-sided counterparts.

\begin{figure*}[!htb]
\small
\begin{gather*}
\mbox{}\hspace{-5mm}\inferrule*[right=\sf Next]
      {\jgrhol{\Delta}{\Sigma}{\Gamma, x_1:A_1, x_2:A_2}{\Psi, \phi'\defsubst{x_1}{x_2}}{t_1}{A_1}{t_2}{A_2}{\phi} \\
        \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\later{A_1}}{u_2}{\later{A_2}}
        {\dsubst{\res\ltag, \res\rtag}{\res\ltag, \res\rtag}{\phi'}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}
              {\nextt{x_1\!\leftarrow\!u_1}{t_1}}{\later{A_1}} 
              {\nextt{x_2\!\leftarrow\!u_2}{t_2}}{\later{A_2}}
              {\triangleright[x_1\!\leftarrow\!u_1,x_2\!\leftarrow\!u_2,
                \res\ltag\!\leftarrow\!\res\ltag,\res\rtag\!\leftarrow\!\res\rtag].\phi}} 
\\[0.3em]
 \inferrule*[right=\sf Prev]
      {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{\later A_1}{t_2}{\later A_2}{\dsubst{\res\ltag,\res\rtag}{\res\ltag,\res\rtag}{\phi}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\prev{t_1}}{A_1}{\prev{t_2}}{A_2}{\phi}} 
\\[0.3em]
 \mbox{}\hspace{-2mm}\inferrule*[Right=\sf Box]
      {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{A_1}{t_2}{A_2}{\phi}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\boxx{t_1}}{\square A_1}{\boxx{t_2}}{\square A_2}
              {\square \phi\defsubst{\letbox{x_1}{\res\ltag}{x_1}}{\letbox{x_2}{\res\rtag}{x_2}}}} 
\\[0.3em]
 \inferrule*[right=\sf LetBox]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\square B_1}{u_2}{\square B_2}{\square \phi\defsubst{\letbox{x_1}{\res\ltag}{x_1}}{\letbox{x_2}{\res\rtag}{x_2}}} \\
       \jgrhol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi'}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\letbox{x_1}{u_1}{t_1}}{A_1}{\letbox{x_2}{u_2}{t_2}}{A_2}{\phi'}} 
\\[0.3em]
  \inferrule*[right=\sf LetConst]
      {B_1,B_2,\phi\ \text{constant} \\ FV(\phi)\cap FV(\Gamma) = \emptyset \\
        \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{B_1}{u_2}{B_2}{\phi} \\
       \jgrhol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi'}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\letconst{x_1}{u_1}{t_1}}{A_1}{\letconst{x_2}{u_2}{t_2}}{A_2}{\phi'}} 
\\[0.3em]
  \inferrule*[right=\sf Fix]
      {\jgrhol{\Delta}{\Sigma}{\Gamma, f_1:\later{A_1}, f_2:\later{A_2}}{\Psi, \dsubst{\res\ltag,\res\rtag}{f_1,f_2}{\phi}}
              {t_1}{A_1}{t_2}{A_2}{\phi}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\fix{f_1}{t_1}}{A_1}{\fix{f_2}{t_2}}{A_2}{\phi}} 
\\[0.3em]
  \mbox{}\hspace{-5.2mm}\inferrule*[Right=\sf Cons]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{A_1}{x_2}{A_2}{\phi_h} \\
       \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{xs_1}{\later{\Str{A_1}}}{xs_2}{\later{\Str{A_2}}}{\phi_t} \\
       \jgholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\forall x_1,x_2, s_1, s_2. \phi_h\defsubst{x_1}{x_2} \Rightarrow
          \phi_t\defsubst{s_1}{s_2} \Rightarrow \phi\defsubst{\cons{x_1}{s_1}}{\cons{x_2}{s_2}}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\cons{x_1}{s_1}}{\Str{A_1}}{\cons{x_2}{s_2}}{\Str{A_2}}{\phi}}
\\[0.3em]
  \inferrule*[right=\sf Head]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_1}{\Str{A_1}}{\phi\defsubst{hd\ \res\ltag}{hd\ \res\rtag}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{hd\ t_1}{A_1}{hd\ t_2}{A_2}{\phi}} 
\\[0.3em]
   \inferrule*[right=\sf Tail]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_2}{\Str{A_2}}{\phi\defsubst{tl\ \res\ltag}{tl\ \res\rtag}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{tl\ t_1}{\later{\Str{A_1}}}{tl\ t_2}{\later{\Str{A_2}}}{\phi}} 
\end{gather*}
\shrinkcaption
\caption{Two-sided rules for Guarded RHOL}\label{fig:s-rhol}
\end{figure*}

\begin{figure*}[!htb]
\small
\begin{gather*}
\inferrule*[right=\sf Next{-}L]
      {\jgrhol{\Delta}{\Sigma}{\Gamma, x_1:B_1}{\Psi, \phi'\subst{\res}{x_1}}{t_1}{A_1}{t_2}{A_2}{\phi} \\
       \jguholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\later{B_1}}{\dsubst{\res}{\res}{\phi'}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}
              {\later[x_1 \ot u_1]{t_1}}{\later{A_1}} 
              {t_2}{A_2}
              {\later[x_1 \ot u_1,\res\ltag\leftarrow\res\ltag]{\phi}}}
\\[0.3em]
  \inferrule*[right=\sf Prev{-}L ]
      {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{\later A_1}{t_2}{A_2}{\dsubst{\res\ltag}{\res\ltag}{\phi}}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma_1; \Gamma_2}{\Psi_1 ; \Psi_2}{\prev{t_1}}{A_1}{t_2}{A_2}{\phi}}
\\[0.3em]
 \inferrule*[right=\sf Box{-}L]
      {\jgrhol{\Delta}{\Sigma}{\Gamma_2}{\Psi_2}{t_1}{A_1}{t_2}{A_2}{\phi} \\ FV(t_1)\not\subseteq FV(\Gamma_2)\\
      FV(\Psi_2)\subseteq FV(\Gamma_2)}
      {\jgrhol{\Delta}{\Sigma}{\Gamma_1; \Gamma_2}{\Psi_1; \Psi_2}{\boxx{t_1}}{\square A_1}{t_2}{A_2}
              {\square \phi\subst{\res\ltag}{\letbox{x_1}{\res\ltag}{x_1}}}} 
\\[0.3em]
 \inferrule*[right=\sf LetBox{-}L]
      {\jguholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\square B_1}{\square \phi\subst{\res}{\letbox{x_1}{\res\ltag}{x_1}}} \\
       \jgrhol{\Delta, x_1 : B_1}{\Sigma, \phi\subst{\res}{x_1}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi'}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\letbox{x_1}{u_1}{t_1}}{A_1}{t_2}{A_2}{\phi'}} 
\\[0.3em]
  \inferrule*[right=\sf LetConst{-}L]
      {B_1,\phi\ \text{constant} \\ FV(\phi)\cap FV(\Gamma) = \emptyset \\
        \jguholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{B_1}{\phi} \\\\
       \jgrhol{\Delta, x_1 : B_1}{\Sigma, \phi\subst{\res}{x_1}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi'}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\letconst{x_1}{u_1}{t_1}}{A_1}{t_2}{A_2}{\phi'}}
\\[0.3em]
  \inferrule*[right=\sf Fix{-}L]
      {\jgrhol{\Delta}{\Sigma}{\Gamma, f_1:\later{A_1}}{\Psi, \dsubst{\res\ltag}{f_1}{(\phi\subst{\res\rtag}{t_2})}}
              {t_1}{A_1}{t_2}{A_2}{\phi}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\fix{f_1}{t_1}}{A_1}{t_2}{A_2}{\phi}}
\\[0.3em]
  \inferrule*[right=\sf Cons{-}L]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{A_1}{t_2}{A_2}{\phi_h} \\
       \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{xs_1}{\later{\Str{A_1}}}{t_2}{A_2}{\phi_t} \\
       \jgholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\forall x_1,x_2, xs_1. \phi_h\defsubst{x_1}{x_2} \Rightarrow
          \phi_t\defsubst{xs_1}{x_2} \Rightarrow \phi\defsubst{\cons{x_1}{xs_1}}{x_2}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\cons{x_1}{xs_1}}{\Str{A_1}}{t_2}{A_2}{\phi}} 
\\[0.3em]
  \inferrule*[right=\sf Head{-}L]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_1}{A_2}{\phi\subst{\res\ltag}{hd\ \res\ltag}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{hd\ t_1}{A_1}{t_2}{A_2}{\phi}}
\\[0.3em]
   \inferrule*[right=\sf Tail{-}L]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_2}{A_2}{\phi\subst{\res\ltag}{tl\ \res\ltag}}}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{tl\ t_1}{\later{\Str{A_1}}}{t_2}{A_2}{\phi}} 
\end{gather*}
\shrinkcaption
\caption{One-sided rules for Guarded RHOL}\label{fig:a-rhol}
\end{figure*}





\subsection{Metatheory}
We review some of the most interesting metatheoretical properties of
our relational proof system, highlighting the equivalence with Guarded HOL.
\begin{theorem}[Equivalence with Guarded HOL] \label{thm:equiv-rhol-hol}
  For all contexts $\Delta,\Gamma$; types $\sigma_1,\sigma_2$; terms $t_1,t_2$; sets of assertions $\Sigma,\Psi,$;
  and assertions $\phi$:
    \[\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}
    \quad\Longleftrightarrow\quad
    \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{\res\ltag}{t_1}\subst{\res\rtag}{t_2}}\]
\end{theorem}
The forward implication follows by induction on the given derivation.
The reverse implication is immediate from the rule which allows to
fall back on Guarded HOL in relational proofs. (Rule \rname{SUB} in the appendix).
The full proof is in the appendix. The consequence of this theorem is
that the syntax-directed, relational proof system we have built 
on top of Guarded HOL does not lose expressiveness.

The intended semantics of a judgement
$\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\phi}$ is
that, for every valuation $\delta \models \Delta$, $\gamma\models \Gamma$,
if $\sem{\Sigma}(\delta)$ and $\sem{\Psi}(\delta,\gamma)$, then
$$\sem{\phi}(\delta,\gamma[\res\ltag \ot \sem{t_1}(\delta,\gamma),
  \res\rtag\ot\sem{t_2}(\delta,\gamma)])$$
Since Guarded HOL is sound with respect to its semantics in
the topos of trees, and our relational proof system is equivalent to
Guarded HOL, we obtain that our relational proof system is also sound in
the topos of trees.

\begin{corollary}[Soundness and consistency]\label{cor:rhol:sound}
If $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_2}{t_2}{\sigma_2}{\phi}$, then for every valuation
$\delta \models \Delta$, $\gamma\models\Gamma$:
\[\begin{array}{c}
  \sem{\Delta \vdash \square \Sigma}(\delta) \wedge \sem{\Delta \mid \Gamma \vdash \Psi}(\delta,\gamma) \Rightarrow
    \\ \sem{\Delta \mid \Gamma,\res\ltag:\sigma_1, \res\ltag:\sigma_2 \vdash \phi}
         (\delta, \gamma[\res\ltag \ot \sem{\Delta \mid \Gamma \vdash t_1}(\delta,\gamma)][\res\rtag \ot \sem{\Delta \mid \Gamma \vdash t_2}(\delta,\gamma)])
\end{array}\]
In particular, there is no proof of
$\jgrhol{\Delta}{\emptyset}{\Gamma}{\emptyset}{t_1}{\sigma_1}{t_2}{\sigma_2}{\bot}$.
\end{corollary}







 \subsection{Shift couplings revisited}
\label{sec:examples}
We give further details on how to prove the example with shift couplings from Section~\ref{sec:proba-ex}.
(Additional examples of relational reasoning on non-probabilistic streams can be
found in the appendix.)
Recall the step functions:$$\begin{array}{rcl}
  \operatorname{step} & \defeq & \lambda x.\mlet{z}{\mathcal{U}_{\{-1,1\}}}{\munit{z+x}} \\
  \operatorname{lstep2} & \defeq & \lambda x.\mlet{z}{\mathcal{U}_{\{-1,1\}}}{\mlet{b}{\mathcal{U}_{\{0,1\}}}{\munit{x+2*z*b}}}
\end{array}$$
We axiomatize the predicate $\All_{2,1}$, which relates the element at position $2i$ in one stream to the element at position $i$ in another stream, as follows.
\[\begin{array}{l}
    \forall x_1 x_2 xs_1 xs_2 y_1. \phi\subst{x_1}{z_1}\subst{x_2}{z_2} \Rightarrow \\
    \quad  \later[ys_1\ot xs_1]{\later[zs_1 \ot ys_1, ys_2 \ot xs_2]{\All_{2,1}(zs_1, ys_2, z_1.z_2.\phi)}} \Rightarrow \\
    \quad\quad   \All_{2,1}(\cons{x_1}{\cons{y_1}{xs_1}},\cons{x_2}{xs_2}, z_1.z_2. \phi)
\end{array}
\]
In fact, we can assume that, in general, we have a family of
$\All_{m_1, m_2}$ predicates relating two streams at positions
$m_1\cdot i$ and $m_2\cdot i$ for every $i$.

We can now express the existence of a shift coupling by the statement:
{\small
\[p_1 = p_2 \vdash \markov(p_1, \operatorname{step}) \sim \markov(p_2,
  \operatorname{lstep2}) \mid
  \diamond_{\left[\substack{ y_1 \ot \res\ltag\\ y_2 \ot \res\rtag}\right]}\All_{2,1}(y_1,y_2,z_1.z_2.z_1=z_2) \]}
For the proof, we need to introduce an asynchronous rule for Markov chains:
{\small\[
\inferrule*[right=\sf Markov-2-1]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{C_1}{t_2}{C_2}{\phi} \\\\
       \begin{array}{c}
         \jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{(\lambda x_1. \mlet{x'_1}{h_1\ x_1}{h_1\ x'_1})}{C_1\to\Distr(C_1)}{h_2}{C_2\to \Distr(C_2)}
              {\\ \forall x_1 x_2. \phi\subst{z_1}{x_1}\subst{z_2}{x_2} \Rightarrow \diamond_{[ z_1 \ot \res\ltag\ x_1, z_2 \ot \res\rtag\ x_2]}\phi}
       \end{array}}
      {\begin{array}{c}\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{\markov(t_1,h_1)}{\Distr(\Str{C_1})}{\markov(t_2,h_2)}{\Distr(\Str{C_2})}
              {\\ \diamond_{[y_1 \ot \res\ltag, y_2 \ot \res\rtag]}\All_{2,1}(y_1,y_2,z_1.z_2.\phi)}
        \end{array}}
\]}
This asynchronous rule for Markov chains shares the motivations of the
rule for loops proposed in~\cite{BartheGHS17}. Note that one can define
a rule \rname{Markov-m-n} for arbitrary $m$ and $n$ to prove a judgement
of the form $\All_{m,n}$ on two Markov chains.




We show the proof of the shift coupling. By equational reasoning, we get:
\[
\begin{array}{rl}
  & \lambda x_1. \mlet{x'_1}{h_1\ x_1}{h_1\ x'_1}\;
  \equiv \;
 \lambda x_1. \mlet{z_1}{\unif{\{-1,1\}}}
              {h_1\ (z_1 + x_1)} \; \equiv \\
  &\equiv \; \lambda x_1. \mlet{z_1}{\unif{\{-1,1\}}}
                {\mlet{z'_1}{\unif{\{-1,1\}}}
                 {\munit{z'_1 + z_1 + x'_1}}}
\end{array}
\]
and the only interesting premise of \rname{Markov-2-1} is:
\[\left.\begin{array}{rl}
   \lambda x_1. &\mlet{z_1}{\unif{\{-1,1\}}}
               {\\ &\mlet{z'_1}{\unif{\{-1,1\}}}
               {\\ &\munit{z'_1 + z_1 + x'_1}}}
  \end{array}
  \sim
  \begin{array}{rl}
  \lambda x_2. &\mlet{z_2}{\unif{\{-1,1\}}}
               {\\ &\mlet{b_2}{\unif{\{1,0\}}}
               {\\ &\munit{x_2+2*b_2*z_2}}}
  \end{array}
  \;\right|\;
  \begin{array}{c}
    \forall x_1 x_2. x_1 = x_2 \Rightarrow \\
    \res\ltag\ x_1 \stackrel{\diamond}{=} \res\rtag\ x_2
  \end{array}\]
Couplings between $z_1$ and $z_2$ and between $z_1'$ and $b_2$ can be found by simple computations.
This completes the proof.
  







 




\section{Related work}
\label{sec:rw}
Our probabilistic guarded $\lambda$-calculus and the associated logic
Guarded HOL build on top of the guarded $\lambda$-calculus and its internal 
logic~\cite{CBGB16}.  The guarded $\lambda$-calculus has been extended
to guarded dependent type theory~\cite{BGCMB16}, which can be
understood as a theory of guarded refinement types and as a foundation
for proof assistants based on guarded type theory. These systems do
not reason about probabilities, and do not support syntax-directed
(relational) reasoning, both of which we support.

Relational models for higher-order programming languages are often
defined using logical relations. \cite{PlotkinA93} showed how to use
second-order logic to define and reason about logical relations for
the second-order lambda calculus.  Recent work has extended this approach
to logical relations for higher-order programming languages with
computational effects such as nontermination, general references, and
concurrency~\cite{DreyerAB11,caresl,ipm,Krogh-Jespersen17}.  The
logics used in \emph{loc. cit.} are related to our work in two ways:
(1) the logics in \emph{loc. cit.} make use of the later modality for
reasoning about recursion, and (2) the models of the logics in
\emph{loc. cit.} can in fact be defined using guarded type theory.
Our work is more closely related to Relational Higher Order
Logic~\cite{ABGGS17}, which applies the idea of logic-enriched type
theories~\cite{AczelG00,AczelG06} to a relational setting. There
exist alternative approaches for reasoning about relational properties
of higher-order programs; for instance,~\cite{GrimmMFHMPRSB17} have
recently proposed to use monadic reification for reducing relational
verification of $F^*$ to proof obligations in higher-order logic.




A series of work develops reasoning methods for probabilistic
higher-order programs for different variations of the lambda calculus.
One line of work has focused on operationally-based techniques for
reasoning about contextual equivalence of programs.  The methods are
based on probabilistic
bisimulations~\cite{DBLP:conf/esop/CrubilleL14,DBLP:conf/popl/SangiorgiV16}
or on logical relations~\cite{DBLP:conf/fossacs/BizjakB15}.  Most of
these approaches have been developed for languages with discrete
distributions, but recently there has also been work on languages with
continuous
distributions~\cite{DBLP:conf/icfp/BorgstromLGS16,DBLP:conf/esop/CulpepperC17}.
Another line of work has focused on denotational models, starting with
the seminal work in~\cite{Jones:powerdomain-evaluations}.  Recent work
includes support for relational reasoning about equivalence of
programs with continuous distributions for a total programming
language~\cite{DBLP:conf/lics/StatonYWHK16}.  Our approach is most
closely related to prior work based on relational refinement types for
higher-order probabilistic programs. These were initially considered
by~\cite{BFGSSZ14} for a stateful fragment of $F^*$, and later
by~\cite{BGGHRS15,BFGGGHS16} for a pure language.  Both systems are
specialized to building probabilistic couplings; however, the latter
support approximate probabilistic couplings, which yield a natural
interpretation of differential privacy~\cite{DR14}, both in its
vanilla and approximate forms (i.e.\, $\epsilon$- and
$(\epsilon,\delta)$-privacy). Technically, approximate couplings are
modelled as a graded monad, where the index of the monad tracks the privacy budget ($\epsilon$ or $(\epsilon,\delta)$).
Both systems are strictly syntax-directed, and cannot reason about
computations that have different types or syntactic structures, while
our system can.





\section{Conclusion}
\label{sec:conclusion}

We have developed a probabilistic extension of the (simply typed)
guarded $\lambda$-calculus, and proposed a syntax-directed proof
system for relational verification. Moreover, we have verified a
series of examples that are beyond the reach of prior work. Finally,
we have proved the soundness of the proof system with respect to the
topos of trees.

There are several natural directions for future work. One first
direction is to enhance the expressiveness of the underlying simply
typed language. For instance, it would be interesting to introduce
clock variables and some type dependency as in~\cite{BGCMB16}, and extend the proof system accordingly.
This would allow us, for example, to type the function taking the $n$-th element of a \emph{guarded} stream,
which cannot be done in the current system.
Another exciting direction is to
consider approximate couplings, as in~\cite{BGGHRS15,BFGGGHS16},
and to develop differential privacy for infinite streams---preliminary
work in this direction, such as~\cite{KellarisPXP14}, considers very
large lists, but not arbitrary streams. A final direction would be to
extend our approach to continuous distributions to support
other application domains.



 
\subsubsection*{Acknowledgments.}
We would like to thank the anonymous reviewers for their time and their helpful input.
This research was supported in part by the ModuRes Sapere Aude Advanced Grant from The Danish Council for Independent Research for the Natural Sciences (FNU),
 by a research grant (12386, Guarded Homotopy Type Theory) from
the VILLUM foundation, and by NSF under grant 1718220.


\bibliography{refs}
\appendix
\newpage


\section{Additional proof rules}

\begin{figure*}[!htb]
\small
\begin{mathpar}
\infer[\sf ABS]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\lambda x_1:\tau_1. t_1}{\tau_1 \to \sigma_1}{\lambda x_2:\tau_2. t_2}{\tau_2\to \sigma_2}{\forall x_1,x_2. \phi' \Rightarrow \phi\subst{\res\ltag}{\res\ltag\ x_1}\subst{\res\rtag}{\res\rtag\ x_2}}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma,x_1:\tau_1,x_2:\tau_2}{\Psi,\phi'}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}}
\and
\infer[\sf APP]{\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1 u_1}{\sigma_1}{t_2 u_2}{\sigma_2}{\phi\subst{x_1}{u_1}\subst{x_2}{u_2}}}
      {\begin{array}{c}
\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\tau_1\to \sigma_1}{t_2}{\tau_2\to \sigma_2}{
          \forall x_1,x_2. \phi'\subst{\res\ltag}{x_1}\subst{\res\rtag}{x_2}\Rightarrow \phi\subst{\res\ltag}{\res\ltag\ x_1}\subst{\res\rtag}{\res\rtag\ x_2}}\\
\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\tau_1}{u_2}{\tau_2}{
          \phi'}
\end{array}}
\and
\infer[\sf VAR]{\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{\sigma_1}{x_2}{\sigma_2}{\phi}}
      {\jlc{\Delta\mid\Gamma}{x_1}{\sigma_1} & \jlc{\Delta\mid\Gamma}{x_2}{\sigma_2} & 
       \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{\res\ltag}{x_1}\subst{\res\rtag}{x_2}}}
\and
\infer[\sf SUB]{\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi'} & 
       \Delta \mid \Sigma \mid \Gamma \mid \Psi \vdash_{\sf GHOL} \phi'\defsubst{t_1}{t_2} \Rightarrow \phi\defsubst{t_1}{t_2}}
\and
\infer[\sf UHOL-L]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{\phi\defsubst{\res}{t_2}}
      \\  \jlc{\Delta\mid\Gamma}{t_2}{\sigma_2}}
\and
\infer[\sf ABS{-}L]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\lambda x_1:\tau_1. t_1}{\tau_1 \to \sigma_1}{t_2}{\sigma_2}{\forall x_1. \phi' \Rightarrow \phi \subst{\res\ltag}{\res\ltag\ x_1}}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma,x_1:\tau_1}{\Psi, \phi'}{t_1}{\sigma_1}{t_2}{\sigma_2}{\phi}}
\and
\infer[\sf APP{-}L]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1 u_1}{\sigma_1}{u_2}{\sigma_2}{\phi\subst{x_1}{u_1}}}
      {\begin{array}{c}
          \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\tau_1\to \sigma_1}{u_2}{\sigma_2}{\forall x_1. \phi'\subst{\res\ltag}{x_1} \Rightarrow \phi\subst{\res\ltag}{\res\ltag\ x_1}}\\
          \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\sigma_1}{\phi'}
\end{array}}
\and
\infer[\sf VAR{-}L]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{\sigma_1}{t_2}{\sigma_2}{\phi}}
      {\phi\subst{\res\ltag}{x_1} \in \Psi &  \res\rtag\not\in\ FV(\phi) &
       \jlc{\Delta\mid\Gamma}{t_2}{\sigma_2}}
\and      
  \inferrule*[Right=\sf Equiv]
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t'_1}{A_1}{t'_2}{A_2}{\Phi} \\ t_1 \equiv t_1' \\ t_2 \equiv t_2' \\
      \Delta \mid \Gamma \vdash t_1 : A_1 \\ \Delta \mid \Gamma \vdash t_2 : A_2}
      {\jgrholsc{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\Phi}}
\end{mathpar}
\caption{Selected RHOL rules}\label{fig:sel-rhol}
\end{figure*}

\section{Denotational semantics}

\subsection{Types and terms in context}

The meaning of terms is given by the denotational model in the category $\trees$ of presheaves over $\omega$, the first infinite ordinal.
This category $\trees$ is also known as the \emph{topos of trees}~\cite{Birkedal-et-al:topos-of-trees}.
In previous work~\cite{CBGB16} it was shown how to model most of the constructions of the guarded lambda calculus and the associated logic, with the notable exception of the probabilistic features.
Below we give an elementary and self-contained presentation of the semantics.

Concretely, objects $X$ of $\trees$ are families of sets $X_i$ indexed over $\nat$ together with functions $r_n^X : X_{n+1} \to X_n$.
These are called \emph{restriction functions}.
We will write simply $r_n$ if $X$ is clear from the context.
Moreover if $x \in X_i$ and $j \leq i$ we will write $x\restriction_j$ for the element $r_j(\cdots (r_{i-1}(x))\cdots) \in X_j$.
Morphisms $X \to Y$ are families of functions $\alpha_n : X_n \to Y_n$ commuting with restriction functions in the sense of $r_n^Y \circ \alpha_{n+1} = \alpha_n \circ r_n^X$.
One can see the restriction function $r_n : X_{n+1} \to X_n$ as mapping elements of $X_{n+1}$ to their approximations at time $n$.

Semantics of types can be found on \autoref{fig:sem-types}, where
$G\left(\sem{A}\right)$ consists of sequences $\{x_n\}_{n \in \nat}$
such that $x_i \in \sem{A}_i$ and $r_i(x_{i+1}) = x_i$ for all $i$,
i.e., $\square\sem{A}$ is the set of so-called global sections of
$\sem{A}$.


The semantics of a dual context $\Delta \mid \Gamma$ is given as the product of types in $\Delta$ and $\Gamma$, except that we implicitly add $\square$ in front of every type in $\Delta$.
In the particular case when both contexts are empty, the semantics of the dual context correspond to the terminal object $1$, which is the singleton set $\{\ast\}$ at each stage. A term in context $\Delta \mid \Gamma \vdash t : \tau$ is interpreted as a family of functions $\sem{t}_n : \sem{\Delta \mid \Gamma}_n \to \sem{\tau}_n$ commuting with restriction functions of $\sem{\Delta \mid \Gamma}$ and $\sem{\tau}$.
Semantics of products,
coproducts, and natural numbers is pointwise as in sets, so we
omit writing it.  The cases for the other constructs are in
\autoref{fig:sem-glc} where $\mathsf{munit}$ and $\mathsf{mlet}$
are the standard unit and bind operations on discrete probabilities,
i.e.\,
\begin{align*}
  \begin{array}{rcl}
    \munit{c} & = & \lambda y.~\carac{c=y} \\
    \mathsf{mlet}~x = \mu~\mathsf{in}~M & = & \lambda y.~\sum_{c\in C} \mu(c) \cdot M(c)(y)
  \end{array}
\end{align*}
The functions $\pi_0$ and $\pi_1$ are the first and second projections, respectively.

\begin{figure*}[!htb]
  
\begin{align*}
  \sem{b} &\defeq \text{ chosen object of } \trees\\
  \sem{\nat} &\defeq \nat \xleftarrow{id} \nat \xleftarrow{id} \nat \xleftarrow{id} \cdots\\
\sem{A\times B} &\defeq \sem{A}_0\times\sem{B}_0 \xleftarrow{r_0 \times r_0} \sem{A}_1\times\sem{B}_1 \xleftarrow{r_1 \times r_1} \sem{A}_2\times\sem{B}_2 \xleftarrow{r_2 \times r_2} \cdots \\
  \sem{A\to B} &\defeq \left(\sem{B}^{\sem{A}}\right)_0 \xleftarrow{\pi} \left(\sem{B}^{\sem{A}}\right)_1 \xleftarrow{\pi} \left(\sem{B}^{\sem{A}}\right)_2 \xleftarrow{\pi} \cdots \\
  \sem{\Str{A}} &\defeq \sem{A}_0 \times \{*\} \xleftarrow{r_0 \times !} \sem{A}_1 \times \left(\sem{A}_0 \times \{*\}\right)
                 \xleftarrow{r_1 \times r_0 \times !} \sem{A}_2\times\left(\sem{A}_1\times\left(\sem{A}_0 \times \{*\}\right)\right) \leftarrow \cdots \\
  \sem{\later A} &\defeq \{*\} \xleftarrow{!} \sem{A}_0 \xleftarrow{r_0} \sem{A}_1 \xleftarrow{r_1} \cdots \\
  \sem{\square A} &\defeq G(\sem{A}) \xleftarrow{id} G(\sem{A}) \xleftarrow{id} \cdots\\
  \sem{\Distr(C)} &\defeq \Distr(\sem{C}_0) \stackrel{\Distr(r_0)}{\longleftarrow} \Distr(\sem{C}_1) \stackrel{\Distr(r_1)}{\longleftarrow} \Distr(\sem{C}_2) \stackrel{\Distr(r_2)}{\longleftarrow} \ldots
\end{align*}
\caption{Semantics of types in the topos of trees}
\label{fig:sem-types}
\end{figure*}

\begin{figure*}[!htb]
\begin{align*}
&\sem{\Delta \mid \Gamma \vdash \lambda x:A. t : A \to B}_i(\delta,\gamma) 
      \defeq (f_0,\ldots,f_i)\\
      &\quad\quad\text{ where } f_i(x) = \sem{\Delta\mid \Gamma, x : A \vdash t : B}\left(\delta,(\gamma\restriction_i, x)\right)\\
  &\sem{\Delta \mid \Gamma \vdash t_1\ t_2 : B}_i(\delta,\gamma)
     \defeq f_i\left(\sem{\Delta \mid\Gamma \vdash t_2 : A}(\delta,\gamma)\right)\\
   &\quad\quad\text{ where } \sem{\Delta \mid \Gamma \vdash t_1 : A \to B}_i(\delta,\gamma) = (f_0,\ldots,f_i)\\
  &\sem{\Delta \mid \Gamma \vdash \later[x_1\ot t_1, \dots, x_n\ot t_n]t : \later A}_0(\delta,\gamma) \defeq *\\ 
  &\sem{\Delta \mid \Gamma \vdash \later[x_1\ot t_1, \dots, x_n\ot t_n]t : \later A}_{i+1}(\delta,\gamma) \defeq \\
     &\quad\quad\sem{\Delta \mid \Gamma, \{x_k : A_k\}_{k=1}^n \vdash t : A}_i\left(\delta, (\gamma\restriction_i, \left\{\sem{\Delta \mid \Gamma \vdash t_k : \later A_k}_{i+1}\right\}_{k=1}^n(\delta,\gamma))\right)\\
  &\sem{\Delta \mid \Gamma \vdash \prev t : A}_i(\delta,\gamma) \defeq \sem{\Delta \mid \cdot \vdash t : \later A}_{i+1}(\delta)\\
  &\sem{\Delta \mid \Gamma \vdash \boxx t : \square A}_i(\delta,\gamma) \defeq \{\sem{\Delta \mid \cdot \vdash t: A}_j(\delta)\}_{j=0}^{\infty}\\
  &\sem{\Delta \mid \Gamma \vdash \letbox{x}{u}{t} : A}_i(\delta,\gamma) \defeq \\
    &\quad\quad\sem{\Delta, x : B \mid \Gamma \vdash t : A}_i\left(\left(\delta,\sem{\Delta \mid \Gamma \vdash u : \square B}_i(\delta,\gamma)\right),\gamma\right)\\
  &\sem{\Delta \mid \Gamma \vdash \letconst{x}{u}{t} : A}_i(\delta,\gamma)\defeq\\
    &\quad\quad\sem{\Delta, x : B \mid \Gamma \vdash t : A}_i\left(\left(\delta,\varepsilon^{-1}_i\left(\sem{\Delta \mid \Gamma \vdash u : \square B}_i(\delta,\gamma)\right)\right),\gamma\right)\\
  &\sem{\Delta \mid \Gamma \vdash \hd t : A}_i(\delta,\gamma) \defeq \pi_0\left(\sem{\Delta \mid \Gamma \vdash t : \Str{A}}_i(\delta,\gamma)\right)\\
  &\sem{\Delta \mid \Gamma \vdash \tl t : \later{\Str{A}}}_i(\delta,\gamma) \defeq \pi_1\left(\sem{\Delta \mid \Gamma \vdash t : \Str{A}}_i(\delta,\gamma)\right)\\
  &\sem{\Delta \mid \Gamma \vdash \cons{t}{u} : \Str{A}}_i(\delta,\gamma) \defeq
   \left(\sem{\Delta \mid \Gamma \vdash t : A}_i(\delta,\gamma), \sem{\Delta \mid \Gamma \vdash u : \Str{A}}_i(\delta,\gamma)\right)\\
  &\sem{\Delta \mid \Gamma \vdash \munit{t} : \Distr(C)}_i (\delta,\gamma)
   \defeq \munit{\sem{\Delta \mid \Gamma \vdash t:C}_i (\delta,\gamma)}  \\
  &\sem{\Delta \mid \Gamma \vdash \mlet{x}{t}{u} : \Distr(C)}_i (\delta,\gamma)
   \defeq
  \mathsf{mlet}~v= \sem{\Delta \mid \Gamma \vdash t : \Distr(D)}_i
  (\delta,\gamma)~\mathsf{in} \\
   \mbox{}\hspace{0.5cm}
    &\quad\quad\sem{\Delta \mid \Gamma,x:D \vdash u : \Distr(C)}_i (\delta,\gamma[x:=v])
\end{align*}
\caption{Semantics for the Guarded $\lambda$-calculus}
\label{fig:sem-glc}
\end{figure*}






\subsection{Equational theory of the calculus}



The denotational semantics validates the following equational theory in addition to the standard equational theory of the simply typed lambda calculus with sums and natural numbers.\\
\textbf{Rules for fixed points, always modality and streams}
\begin{align*}
  \begin{split}
\fix{f}{t} \quad &\equiv \quad t\subst{f}{\later(\fix{f}{t})} \\
    \prev{(\later t)} \quad &\equiv \quad t \\
    \letbox{x}{(\boxx{u})}{t} \quad &\equiv \quad t\subst{x}{u} \\
    \letconst{x}{u}{t} \quad&\equiv\quad t\subst{x}{u} \\
  \end{split}
  \begin{split}
    \hd\ (\cons{x}{xs}) \quad&\equiv\quad x \\
    \tl\ (\cons{x}{xs}) \quad&\equiv\quad xs \\
    \cons{\hd t}{\tl t} \quad&\equiv\quad t
  \end{split}
\end{align*}
\textbf{Rules for delayed substitutions}
\begin{align*}
  \latern[\xi\hrt{x\gets t}]{u} &\equiv \latern[\xi]{u} & & \text{ if } x \text{ not in } u\\
  \latern[\xi\hrt{x\gets t, y \gets s}\xi']{u} &\equiv\latern[\xi\hrt{y\gets s, x \gets t}\xi']{u}\\
  \latern[\xi\hrt{x\gets \latern[\xi]{t}}]{u} &\equiv \latern[\xi]{\left(u\subst{x}{t}\right)}\\
  \latern[\hrt{x\gets t}]x &\equiv t
\end{align*}
\textbf{Monad laws for distributions}
\begin{align*}
  \mlet{x}{\munit{t}}{u} &\equiv u[t/x]\\
  \mlet{x}{t}{\munit{x}} &\equiv t\\
  \mlet{x_2}{(\mlet{x_1}{t_1}{t_2})}{u} &\equiv \mlet{x_1}{t_1}{(\mlet{x_2}{t_2}{u})}
\end{align*}
In particular, notice that fix does not reduce as usual, but instead the whole term is delayed before the substitution is performed.



\subsection{Logical judgements}


The cases for the semantics of the judgement $\Delta \mid \Gamma \vdash \phi$ of the non-probabilistic fragment are as follows (we omit writing the contexts if they are clear):
\begin{align*}
  \sem{\top}_i &\defeq \sem{\Delta \mid \Gamma}_i\\
  \sem{\phi \wedge \psi}_i &\defeq \sem{\phi}_i \cap \sem{\psi}_i\\
  \sem{\phi \vee \psi}_i &\defeq \sem{\phi}_i \cup \sem{\psi}_i\\
  \sem{\phi \To \psi}_i &\defeq
  \left\{ x \isetsep \forall j \leq i, x\restriction_j \in \sem{\phi}_j \To x\restriction_j \in \sem{\psi}_j\right\}\\
  \sem{\forall x:A. \phi}_i &\defeq \left\{(\delta,\gamma) \isetsep \forall j \leq i, \forall x \in \sem{A}_j, \left(\delta, \left(\gamma\restriction_j\right), x\right) \in \sem{\phi}\right\}\\
  \sem{\later[x_1 \ot t_1, \dots, x_n \ot t_n]\phi}_i &\defeq 
    \left\{ (\delta, \gamma) \isetsep i > 0 \To \left(\delta,\gamma\restriction_{i-1}, \left\{\sem{t_k}_{i}(\delta,\gamma)\right\}_{k=1}^n\right) \in \sem{\phi}_{i-1}\right\}\\
  \sem{\square \phi}_i &\defeq \left\{ x \isetsep \forall j, x \in \sem{\phi}_j\right\}
\end{align*}

\section{Additional background}


One consequence of Strassen's theorem is
that couplings are closed under convex combinations.
\begin{lemma}[Convex combinations of couplings]\label{lem:convex-combination-of-couplings}
Let $(\mu_i)_{i\in I}$ and $(\nu_i)_{i\in I}$ bet two families of
distributions on $C_1$ and $C_2$ respectively, and let $(p_i)_{i\in I}
\in [0,1]$ such that $\sum_{i\in I} p_i=1$. If
$\coupl{\mu_i}{\nu_i}{R}$ for all $i\in I$ then $\coupl{(\sum_{i\in I} p_i
\mu_i)}{(\sum_{i\in I} p_i \nu_i)}{R}$, where the convex
combination $\sum_{i\in I} p_i \mu_i$ is defined by the clause
$(\sum_{i\in I} p_i \mu_i)(x)=\sum_{i\in I} p_i \mu_i(x)$.
\end{lemma}
One obtains an asymmetric version of the lemma by observing that if
$\mu_i=\mu$ for every $i\in I$, then $(\sum_{i\in I} p_i \mu_i)=\mu$.


One can also show that couplings are closed under relation
composition.
\begin{lemma}[Couplings for relation composition]
Let $\mu_1\in\Distr(C_1)$, $\mu_2\in\Distr(C_2)$, $\mu_3
\in\Distr(C_3)$. Moreover, let $R\subseteq C_1\times C_2$ and
$S\subseteq C_2\times C_3$. If $\coupl{\mu_1}{\mu_2}{R}$ and
$\coupl{\mu_2}{\mu_3}{R}$ then $\coupl{\mu_1}{\mu_3}{R}$.
\end{lemma}

\section{Proofs of the theorems}


\subsection{Proof of Theorem~\ref{thm:sound-ghol}}
\label{sec:soundness-of-ghol}

The semantics of the guarded higher-order logic without the probabilistic fragment has been explained in previous work~\cite{Birkedal-et-al:topos-of-trees,CBGB16}.
Thus we focus on showing soundness of the additional rules for the diamond modality, which will be useful for proving soundness of the relational proof system.
Moreover we only describe soundness for the binary diamond modality, the soundness of the rules for the unary modality being entirely analogous.


\noindent\textbf{Soundness of the rule \textsf{MONO2}}
\begin{mathpar}
  \infer[\sf MONO2]
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\gets t_1, x_2\gets t_2]}\psi}}
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\gets t_1, x_2 \gets t_2]}\phi} \and
    \jghol{\Delta}{\Sigma}{\Gamma, x_1 : C_1, x_2 : C_2}{\Psi, \phi}{\psi}
  }
\end{mathpar}

Let $n \in \nat$ and $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$.
Then from the first premise we have $(\delta, \gamma) \in \sem{\diamond_{[x_1\gets t_1, x_2 \gets t_2]}\phi}_n$ and thus there exists an $\left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}$ coupling for the distributions $\sem{t_1}_n(\delta,\gamma)$ and $\sem{t_2}_n(\delta,\gamma)$.
But since $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$ we have from the second premise of the rule that
\begin{align*}
  \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\} \subseteq
  \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\psi}_n\right\}
\end{align*}
and thus any $\left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}$ coupling is also an $\left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\psi}_n\right\}$ coupling, which means there exists an $\left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\psi}_n\right\}$ coupling for $\sem{t_1}_n(\delta,\gamma)$ and $\sem{t_2}_n(\delta,\gamma)$ as required.

\noindent\textbf{Soundness of the rule \textsf{UNIT2}}
\begin{mathpar}
  \infer[\sf UNIT2]
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \diamond_{[x_1\leftarrow \munit{t_1}, x_2\leftarrow \munit{t_2}]} \phi}}
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{ \phi[t_1/x_1][t_2/x_2]}}
\end{mathpar}
Let $n \in \nat$ and $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$.
We need to show the existence of a
\begin{align*}
  \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}
\end{align*}
coupling for the point-mass distributions concentrated at $\sem{t_1}_n(\delta,\gamma)$ and $\sem{t_2}_n(\delta,\gamma)$.
The premise of the rule establishes the membership
\begin{align*}
  \left(\sem{t_1}_n(\delta,\gamma), \sem{t_2}_n(\delta,\gamma)\right)
  \in
  \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}
\end{align*}
and thus the point-mass distribution concentrated at $\left(\sem{t_1}_n(\delta,\gamma), \sem{t_2}_n(\delta,\gamma)\right)$ is a required coupling.


\noindent\textbf{Soundess of the rule \textsf{MLET2}}
\begin{mathpar}
\infer[\sf MLET2]
      {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{
          \diamond_{[y_1\leftarrow \mlet{x_1}{t_1}{t'_1}, y_2\leftarrow
              \mlet{x_2}{t_2}{t'_2}]} \psi}}{
          \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1, x_2\leftarrow t_2]} \phi}
          \and
          \jghol{\Delta}{\Sigma}{\Gamma,x_1:C_1,x_2:C_2}{\Psi,\phi}{
            \diamond_{[y_1\leftarrow t'_1, y_2\leftarrow t'_2]}\psi}}
\end{mathpar}
Let $n \in \nat$ and $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$.
Then from the first premise we have that there exists an
\begin{align*}
  \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}
\end{align*}
coupling for the distributions $\sem{t_1}_n(\delta,\gamma)$ and $\sem{t_2}_n(\delta,\gamma)$.

From the second premise we get that for every $v, u$ such that $(\delta, \gamma, v, u) \in \sem{\phi}_n$
there exists an 
\begin{align*}
  \left\{(v', u') \isetsep (\delta, \gamma, v, u, v', u') \in \sem{\psi}_n\right\}
\end{align*}
coupling for $\sem{t_1'}_n(\delta,\gamma, v)$ and $\sem{t_2'}_n(\delta,\gamma, u)$.
Since $x_1$ and $x_2$ are fresh for $\psi$ the relation
\begin{align*}
  \left\{(v', u') \isetsep (\delta, \gamma, v, u, v', u') \in \sem{\psi}_n\right\}
\end{align*}
is independent of $v, u$.

Thus Lemma~\ref{lem:sequential-composition-of-couplings} instantiated with
\begin{align*}
  \mu_1 &= \sem{t_1}_n(\delta,\gamma)\\
  \mu_2 &= \sem{t_2}_n(\delta,\gamma)\\
  M_1 &= \sem{t_1'}_n(\delta,\gamma, -)\\
  M_2 &= \sem{t_2'}_n(\delta,\gamma, -)\\
  R &= \left\{(v, u) \isetsep (\delta, \gamma, v, u) \in \sem{\phi}_n\right\}\\
  S &= \left\{(v', u') \isetsep (\delta, \gamma, v, u, v', u') \in \sem{\psi}_n\right\}
\end{align*}
concludes the proof.

\noindent\textbf{Soundness of the rule \textsf{MLET-L}}
\begin{mathpar}
\infer[\sf MLET-L]
      {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{
          \diamond_{[y_1\leftarrow \mlet{x_1}{t_1}{t'_1}, y_2\leftarrow t'_2]} \psi}}{
          \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x_1\leftarrow t_1]}\phi}
          \and
          \jghol{\Delta}{\Sigma}{\Gamma,x_1:C_1}{\Psi,\phi}{
            \diamond_{[y_1\leftarrow t'_1, y_2\leftarrow t'_2]}\psi}}
\end{mathpar}
Let $n \in \nat$ and $(\delta, \gamma) \in \sem{\Delta \mid \Sigma \mid \Gamma \mid \Psi}_n$.
Then from the first premise we have that the support of the distribution $\sem{t_1}_n(\delta,\gamma)$ is included in
\begin{align*}
  \left\{v \isetsep (\delta, \gamma, v) \in \sem{\phi}_n\right\}.
\end{align*}

From the second premise we get that for every $v$ such that $(\delta, \gamma, v) \in \sem{\phi}_n$ there exists an
\begin{align*}
  \left\{(v', u') \isetsep (\delta, \gamma, v, v', u') \in \sem{\psi}_n\right\}
\end{align*}
coupling for $\sem{t_1'}_n(\delta,\gamma, v)$ and $\sem{t_2'}_n(\delta,\gamma)$.
Since $x_1$ is fresh for $\psi$ the relation
\begin{align*}
  R \defeq \left\{(v', u') \isetsep (\delta, \gamma, v, v', u') \in \sem{\psi}_n\right\}
\end{align*}
is independent of $v$.

Let $\mathcal{I} = \left\{v \isetsep (\delta, \gamma, v) \in \sem{\phi}_n\right\}$ and for any $v \in \mathcal{I}$ let $p_v = \sem{t_1}_n(\delta,\gamma)(v)$, $\mu_v = \sem{t_1'}_n(\delta,\gamma,v)$, and $\nu_v = \sem{t_2'}_n(\delta,\gamma)$.
Then we have $\sum_{v \in \mathcal{I}} p_v = 1$ from the first premise of the rule and $\coupling{R}{\mu_v}{\nu_v}$ for all $v \in \mathcal{I}$ from the second premise.
Lemma~\ref{lem:convex-combination-of-couplings} concludes the proof.
 



\subsection{Proof of Theorem \ref{thm:equiv-rhol-hol}}

The inverse implication follows immediately from the \rname{SUB} rule and
the fact that we can always prove a judgement of the shape
\[ \jgrhol{\Gamma}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\top} \]
for well-typed $t_1$ and $t_2$.

We will prove the direct implication by induction on the
derivation. We will just prove the two-sided rules.  The proofs for the one
sided rule are similar.
\\ \\
\noindent {\bf Case.} $\small\inferrule*[Right=\sf Next]
      {\jgrhol{\Delta}{\Sigma}{\Gamma, x_1:A_1, x_2:A_2}{\Psi, \Phi'\defsubst{x_1}{x_2}}{t_1}{A_1}{t_2}{A_2}{\Phi} \\
       \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\later{A_1}}{u_2}{\later{A_2}}{\dsubst{\res\ltag, \res\rtag}{\res\ltag, \res\rtag}{\Phi'}}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}
              {\nextt{x_1 \leftarrow u_1}{t_1}}{\later{A_1}} 
              {\nextt{x_2 \leftarrow u_2}{t_2}}{\later{A_2}}
              {\triangleright[x_1\leftarrow u_1,x_2\leftarrow x_2,\res\ltag\leftarrow\res\ltag,\res\rtag\leftarrow\res\rtag]\Phi}}$
\\ \\
\noindent
By I.H. $\jghol{\Delta}{\Sigma}{\Gamma, x_1:A_1, x_2:A_n}{\Psi, \Phi'\defsubst{x_1}{x_2}}{\Phi\defsubst{t_1}{t_2}}$, \hfill{}(H1)
\\
and $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[\res\ltag \ot u_1, \res\rtag \ot u_2]{\Phi'}}$ \hfill{}(H2)
\\
To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot u_1, x_2 \ot u_2, \res\ltag \ot \nextt{x_1 \leftarrow u_1}{t_1} , \res\rtag\ot \nextt{x_2 \leftarrow u_2}{t_2}]{\Phi}}$. \hfill{}(G)
\\
By \rname{CONV} we can change the goal (G) into
\\
$\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot u_1, x_2 \ot u_2]{\Phi\defsubst{t_1}{t_2}}}$ \hfill{}(G')
\\
and (H2) into:
\\
$\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\later[x_1 \ot u_1, x_2 \ot u_2]{\Phi'\defsubst{x_1}{x_2}}}$ \hfill{}(H2')
\\
Finally, by applying \rname{$\later_{App}$} to (H1) and (H2) we get (G')
\\



\noindent {\bf Case.} $\small \inferrule*[Right=\sf Prev]
      {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{\later A_1}{t_2}{\later A_2}{\dsubst{\res\ltag,\res\rtag}{\res\ltag,\res\rtag}{\Phi}}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\prev{t_1}}{A_1}{\prev{t_2}}{A_2}{\Phi}}$
\\ \\
\noindent
We just apply \rname{$\later_E$}.
\\


\noindent {\bf Case.} $\small\inferrule*[Right=\sf Box]
      {\jgrhol{\Delta}{\Sigma}{\cdot}{\cdot}{t_1}{A_1}{t_2}{A_2}{\Phi}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\boxx{t_1}}{\square A_1}{\boxx{t_2}}{\square A_2}
              {\square \Phi\defsubst{\letbox{x_1}{\res\ltag}{x_1}}{\letbox{x_2}{\res\rtag}{x_2}}}}$
\\ \\
\noindent
By I.H. $\jghol{\Delta}{\Sigma}{\cdot}{\cdot}{\Phi\defsubst{t_1}{t_2}}$
\\
To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square \Phi\defsubst{\letbox{x_1}{\boxx{t_1}}{x_1}}{\letbox{x_2}{\boxx{t_2}}{x_2}}}$
\\
By \rname{CONV} we can change the goal into:
\\
$\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square \Phi\defsubst{t_1}{t_2}}$
\\
And then we can prove it by \rname{$\square_I$}.
\\

\noindent {\bf Case.}$\small\inferrule*[Right=\sf LetBox]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{\square B_1}{u_2}{\square B_2}{\square \Phi\defsubst{\letbox{x_1}{\res\ltag}{x_1}}{\letbox{x_2}{\res\rtag}{x_2}}} \\
       \jgrhol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \Phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\Phi'}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\letbox{x_1}{u_1}{t_1}}{A_1}{\letbox{x_2}{u_2}{t_2}}{A_2}{\Phi'}}$
\\ \\
\noindent
By I.H. $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square \Phi\defsubst{\letbox{x_1}{u_1}{x_1}}{\letbox{x_2}{u_2}{x_2}}}$ \hfill{}(H1)
\\
and $\jghol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \Phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{\Phi'\defsubst{t_1}{t_2}}$. \hfill{}(H2)
\\
To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\Phi'\defsubst{\letbox{x_1}{u_1}{t_1}}{\letbox{x_2}{u_2}{t_2}}}$ \hfill{}(G)
\\
We instantiate (H2) into:
\[\begin{array}{c}
  \jghol{\Delta}{\Sigma, \Phi\defsubst{\letbox{x_1}{u_1}{x_1}}{\letbox{x_2}{u_2}{x_2}}}{\Gamma}{\Psi}
        {\\ \Phi'\defsubst{t_1\subst{x_1}{\letbox{x_1}{u_1}{x_1}}}{t_2\subst{x_2}{\letbox{x_2}{u_2}{x_2}}}}
        \end{array}\]
And by the equality $t\subst{x}{\letbox{x}{u}{x}} \equiv \letbox{x}{u}{t}$, we get:
\[\begin{array}{c}
   \jghol{\Delta}{\Sigma, \Phi\defsubst{\letbox{x_1}{u_1}{x_1}}{\letbox{x_2}{u_2}{x_2}}}{\Gamma}{\Psi}
         {\\ \Phi'\defsubst{\letbox{x_1}{u_1}{t_1}}{\letbox{x_2}{u_2}{t_2}}} \end{array} \]
and then, by applying \rname{$\square_E$} to (H1) and the previous judgement we get (G).
\\

\noindent {\bf Case.} $\small \inferrule*[Right=\sf LetConst]
      {B_1,B_2,\Phi\ \text{constant} \\ FV(\Phi)\cap FV(\Gamma) = \emptyset \\
        \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u_1}{B_1}{u_2}{B_2}{\Phi} \\
       \jgrhol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \Phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\Phi'}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\letconst{x_1}{u_1}{t_1}}{A_1}{\letconst{x_2}{u_2}{t_2}}{A_2}{\Phi'}}$
\\ \\
\noindent
By I.H. $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\Phi\subst{u_1}{u_2}}$, \hfill{}(H1)
\\
and $\jghol{\Delta, x_1 : B_1, x_2 : B_2}{\Sigma, \Phi\defsubst{x_1}{x_2}}{\Gamma}{\Psi}{\Phi'\defsubst{t_1}{t_2}}$. \hfill{}(H2)
\\
To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\Phi'\defsubst{\letconst{x_1}{u_1}{t_1}}{\letconst{x_2}{u_2}{t_2}}}$ \hfill{}(G)
\\
From (H1) and the fact that $\Phi, B_1$ and $B_2$ are constant, we get:
\\
$\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\square\Phi\defsubst{u_1}{u_2}}$
\\
The rest of the prove is analogous to the previous case.
\\


\noindent {\bf Case.} $\small \inferrule*[Right=\sf Fix]
      {\jgrhol{\Delta}{\Sigma}{\Gamma, f_1:\later{A_1}, f_2:\later{A_2}}{\Psi, \dsubst{\res\ltag,\res\rtag}{f_1,f_2}{\Phi}}
              {t_1}{A_1}{t_2}{A_2}{\Phi}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{fix f_1. t_1}{A_1}{fix f_2. t_2}{A_2}{\Phi}}$
\\ \\
\noindent
By I.H. $\jghol{\Delta}{\Sigma}{\Gamma, f_1:\later{A_1}, f_2:\later{A_2}}{\Psi,\later[\res\ltag\ot f_1, \res\rtag\ot f_2]{\Phi}}{\Phi\defsubst{t_1}{t_2}}$.
\\
To show: $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\Phi\subst{\fix{f_1}{t_1}}{fix{f_2}{t_2}}}$ \\
Instantiating the I.H. with $f_1 = \later \fix{f_1}{t_1}$ and $f_2 = \later \fix{f_2}{t_2}$ we get:
\\
$\jghol{\Delta}{\Sigma}{\Gamma}{\Psi,\later[\res\ltag \ot \later \fix{f_1}{t_1}, \res\rtag \ot \later \fix{f_2}{t_2}]{\Phi}}
       {\Phi\defsubst{t_1\subst{f_1}{\later \fix{f_1}{t_1}}}{t_2\subst{f_2}{\later \fix{f_2}{t_2}}}}$.
\\
Since $t\subst{f}{\later \fix{f}{t}} \equiv \fix{f}{t}$, by \rname{CONV}:
\\
$\jghol{\Delta}{\Sigma}{\Gamma}{\Psi,\later[\res\ltag \ot \later \fix{f_1}{t_1}, \res\rtag \ot \later \fix{f_2}{t_2}]{\Phi}}{\Phi\defsubst{\fix{f_1}{t_1}}{\fix{f_2}{t_2}}}$.
\\
and since $\later[\res\ltag \ot \later \fix{f_1}{t_1}, \res\rtag \ot \later \fix{f_2}{t_2}]{\Phi} \Leftrightarrow \later \Phi\defsubst{\fix{f_2}{t_2}}{\fix{f_2}{t_2}}$,
\\
$\jghol{\Delta}{\Sigma}{\Gamma}{\Psi,\later\Phi\defsubst{\fix{f_1}{t_1}}{fix{f_2}{t_2}}}{\Phi\defsubst{fix{f_1}{t_1}}{fix{f_2}{t_2}}}$,
\\
and finally, by \rname{L\"ob} we get our goal.
\\


\noindent {\bf Case.} $\small\inferrule*[Right=\sf Cons]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x_1}{A_1}{x_2}{A_2}{\Phi_h} \\
       \jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{xs_1}{\later{\Str{A_1}}}{xs_2}{\later{\Str{A_2}}}{\Phi_t} \\
       \jhol{\Gamma}{\Psi}{\forall x_1,x_2, xs_1, xs_2. \Phi_h\defsubst{x_1}{x_2} \Rightarrow
          \Phi_t\defsubst{xs_1}{xs_2} \Rightarrow \Phi\defsubst{\cons{x_1}{xs_1}}{\cons{x_2}{xs_2}}}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\cons{x_1}{xs_1}}{\Str{A_1}}{\cons{x_2}{xs_2}}{\Str{A_2}}{\Phi}}$
\\ \\
Apply the I.H., \rname{$\forall_E$} and \rname{$\To_E$}.
\\






\noindent {\bf Case.}$\small\inferrule*[Right=\sf Head]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_1}{\Str{A_1}}{\Phi\defsubst{hd\ \res\ltag}{hd\ \res\rtag}}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{hd\ t_1}{A_1}{hd\ t_2}{A_2}{\Phi}}$
\\ \\
Trivial by I.H.
\\


\noindent {\bf Case.} $\small\inferrule*[Right=\sf Tail]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\Str{A_1}}{t_2}{\Str{A_2}}{\Phi\defsubst{tl\ \res\ltag}{tl\ \res\rtag}}}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{tl\ t_1}{\later{Str_{A_1}}}{tl\ t_2}{\later{Str_{A_2}}}{\Phi}}$
\\ \\
Trivial by I.H.
\\
  

\noindent {\bf Case.} $\small\inferrule*[Right=\sf Equiv]
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t'_1}{A_1}{t'_2}{A_2}{\Phi} \\ t_1 \equiv t_1' \\ t_2 \equiv t_2' \\
      \Delta \mid \Gamma \vdash t_1 : A_1 \\ \Delta \mid \Gamma \vdash t_2 : A_2}
      {\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{A_1}{t_2}{A_2}{\Phi}}$
\\ \\
Trivial by I.H. and \rname{Conv}.
\\



Most of the proofs for the probabilistic fragment are a consequence of the proof of \autoref{thm:sound-ghol}.
The only interesting case is \rname{Markov}. We do the proof directly in RHOL by showing we can derive it from \rname{Fix}.
We have the premises:
\begin{enumerate}

\item $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{C_1}{t_2}{C_2}{\phi}$

\item $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{h_1}{C_1 \to \Distr(C_1)}{h_2}{C_2 \to \Distr(C_2)}{\psi_3}$

\item $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\psi_4}$
\end{enumerate}

where:
\begin{align*}
\psi_3 &\equiv \forall x_1 x_2. \phi\defsubst{x_1}{x_2} \Rightarrow \diamond_{[ y_1 \ot \res\ltag\ x_1, y_2 \ot \res\rtag\ x_2]}\phi\defsubst{y_1}{y_2} \\
\psi_4 &\equiv \forall x_1\ x_2\ xs_1\ xs_2. \phi\defsubst{x_1}{x_2} \Rightarrow \later[y_1 \ot xs_1, y_2 \ot xs_2]{\Phi}
       \Rightarrow \Phi\subst{y_1}{\cons{x_1}{xs_1}}\subst{y_2}{\cons{x_2}{xs_2}}
\end{align*}

If we inline the definition of unfold, we have to prove:

\[\begin{aligned}[t]
\fix{f}{&\lambda x_1. \lambda h_1. \mlet{z_1}{h_1\ x_1}
                                                   { \mlet{t_1}{\operatorname{swap_{\later\Distr}^C}(f_1\app \later z_1 \app \later h_1)}
                                                   { \munit{\cons{x_1}{t_1}}}}
                                      }\sim \\ \fix{f}{&\lambda x_2. \lambda h_2. \mlet{z_2}{h_2\ x_2}
                                                   { \mlet{t_2}{\operatorname{swap_{\later\Distr}^C}(f_2\app  \later z_2 \app \later h_2)}
                                                   { \munit{\cons{x_2}{t_2}}}}
                                      }\\\mid\; &\forall x_1 x_2 h_1 h_2. \phi\defsubst{x_1}{x_2} \Rightarrow \psi_3\defsubst{h_1}{h_2} \Rightarrow \diamond_{[y_2 \ot\res\ltag, y_2 \ot\res\rtag]}\Phi
\end{aligned}
\]

We apply \rname{FIX}, \rname{MLET} twice, and then \rname{MUNIT}. The main judgements we have to prove are: 

\begin{enumerate}[label*=(\alph*)]

\item $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{h_1\ x_1}{\Distr(C_1)}{h_2\ x_2}{\Distr(C_2)}
              {\diamond_{[\res\ltag\ot\res\ltag, \res\rtag\ot\res\rtag]}\phi}$


\item $\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\operatorname{swap_{\later\Distr}^C}(f_1\app  \later z_1 \app \later h_1)}{\Distr(\later C_1)}
              {\operatorname{swap_{\later\Distr}^C}(f_2\app \later z_2 \app \later h_2)}{\Distr(\later C_2)}
              {\diamond_{[z_1 \ot \res\ltag, z_2 \ot \res\rtag]} \later[y_1 \ot z_1, y_2 \ot z_2]{\Phi}}$

\item $\jgrhol{\Delta}{\Sigma}{\Gamma, x_1,x_2,t_1,t_2}{\Psi, \phi\defsubst{x_1}{x_2}, \later[y_1 \ot t_1, y_2 \ot t_2]{\Phi}}{\cons{y_1}{t_1}}{\Distr(\Str{C_1})}{\cons{y_2}{t_2}}{\Distr(\Str{C_2})}
              {\Phi\subst{y_1}{\res\ltag}\subst{y_2}{\res\rtag}}$
\end{enumerate}

The judgement (a) is a direct consequence of premises (1) and (2), (b) is proven from the inductive hypothesis,
and (d) is a direct consequence of (3). This completes the proof.


 
\section{Examples}


\subsection{Proof of ZipWith}

This example, taken from \cite{BGCMB16}, proves a
property about the $\operatorname{ZipWith}$ function, which takes two
streams of type A, a function on pairs of elements, and ``zips'' the
two streams by applying that function to the elements that are at the
same position on the two streams.  We want to show that if the
function on the elements is commutative, zipping two streams with that
function is commutative as well.

We can define the zipWith function as:
\begin{align*}
  \operatorname{zipWith} &: (\nat \to \nat \to \nat) \to \Str{\nat} \to \Str{\nat} \to \Str{\nat} \\
  \operatorname{zipWith} &\defeq \fix{\operatorname{zipWith}}{\lambda f. \lambda xs. \lambda ys. \cons{(f\ (\hd\ xs)\ (\hd\ ys))}{(\operatorname{zipWith} \app (\tl\ xs) \app (\tl\ ys))}}
\end{align*}
We prove (omitting types of expressions):
$$\jgrholnocnot{\operatorname{zipWith}}{\operatorname{zipWith}}{\Phi}$$
where
\begin{align*}
 \Phi &\defeq \forall f_1 f_2. (f_1 = f_2 \wedge \forall x y. f_1 x y = f_1 y x) \Rightarrow
 \forall xs_1 xs_2. \forall ys_1 ys_2. (xs_1 = ys_2 \wedge xs_2 = ys_1)\\
 & \Rightarrow
                  \res\ltag\ f_1\ xs_1\ ys_1 = \res\rtag\ f_2\ xs_2\ ys_2 \\
 A    &\defeq  (\nat \to \nat \to \nat) \to \Str{\nat} \to \Str{\nat} \to \Str{\nat}
\end{align*}
The proof proceeds by applying two-sided rules all the way. We invite
interested readers to compare this proof with the one given
in~\cite{BGCMB16} to see how the two approaches differ.

We show how to derive the statement backwards. The derivation begins
with the \rname{Fix} rule. Its premise is (omitting constant contexts):
\[
\jrhol{\operatorname{zipWith}_1,\operatorname{zipWith}_2:\later A}{\later[\res\ltag \leftarrow \operatorname{zipWith}_1,\res\rtag \leftarrow \operatorname{zipWith}_2] \Phi}{\lambda f_1.(\cdots)}{A}
      {\lambda f_2.(\cdots)}{A}{\Phi}
      \]
Then we apply the \rname{ABS} rule three times to introduce into the
context the logical relations on $f_1,f_2,xs_1,xs_2,ys_1$, and $ys_2$. The premise
we need to prove is then:
\[
\begin{array}{c}
\jrhol{\Gamma}{\Psi}{\cons{(f_1\ (hd\ xs_1)\ (hd\ ys_1))}{(\operatorname{zipWith}_1 \app (tl\ xs_1) \app (tl\ ys_1))}}{\Str{\nat}}
      {\\ \cons{(f_2\ (hd\ xs_2)\ (hd\ ys_2))}{(\operatorname{zipWith}_2 \app (tl\ xs_2) \app (tl\ ys_2))}}{\Str{\nat}}{\res\ltag = \res\rtag}
\end{array}
\]
where
\begin{align*}
  \Gamma &\defeq \operatorname{zipWith}_1,\operatorname{zipWith}_2 :\later A; f_1,f_2:(\nat\to\nat\to\nat); xs_1,xs_2,ys_1,ys_2: \Str{\nat} \\
  \Psi   &\defeq \later[\res\ltag \leftarrow \operatorname{zipWith}_1,\res\rtag \leftarrow \operatorname{zipWith}_2] \Phi, (f_1 = f_2 \wedge \forall x y. f x y = f y x), xs_1 = ys_2, xs_2 = ys_1
\end{align*}
Now we can apply the \rname{Cons} rule, which has three premises:
\begin{enumerate}
 \item $\jrhol{\Gamma}{\Psi}{f_1\ (hd\ xs_1)\ (hd\ ys_1)}{\nat}
       {f_2\ (hd\ xs_2)\ (hd\ ys_2)}{\nat}{\res\ltag = \res\rtag}$

 \item $\jrhol{\Gamma}{\Psi}{\operatorname{zipWith}_1 \app (tl\ xs_1) \app (tl\ ys_1)}{\later \Str{\nat}}
       {\operatorname{zipWith}_2 \app (tl\ xs_2) \app (tl\ ys_2)}{\later \Str{\nat}}{\res\ltag = \res\rtag}$

 \item $\jhol{\Gamma}{\Psi}{\forall x y xs ys. x=y \Rightarrow xs = ys \Rightarrow \cons{x}{xs} = \cons{y}{ys}}$
\end{enumerate}
Premise (3) is easily provable in HOL. To prove premise (1) we first apply the \rname{App} rule twice, and we have to prove the judgments:
\begin{itemize}
  \item $\jrhol{\Gamma}{\Psi}{f_1}{\nat\to\nat\to\nat}
       {f_2}{\nat\to\nat\to\nat}{ \forall vs_1 vs_2 ws_1 ws_2.
                                   \\ vs_1 = hd\ ys_2 \wedge vs_2 = hd\ ys_1 \Rightarrow
                                     ws_1 = hd\ xs_2 \wedge ws_2 = hd\ xs_1 \Rightarrow \res\ltag\ vs_1\ ws_1 = \res\rtag\ vs_2\ ws_2}$

  \item $\jrhol{\Gamma}{\Psi}{hd\ xs_1}{\nat}
       {hd\ xs_2}{\nat}{\res\ltag = hd\ ys_2 \wedge \res\rtag = hd\ ys_1}$

  \item $\jrhol{\Gamma}{\Psi}{hd\ ys_1}{\nat}
       {hd\ ys_2}{\nat}{\res\ltag = hd\ xs_2 \wedge \res\rtag = hd\ xs_1}$
\end{itemize}
The three can be proven in HOL from the conditions imposed on $f_1,f_2$ and the equalities $xs_1 = ys_2$, $xs_2 = ys_1$.

All that remains to prove is premise (2) of the \rname{Cons} application, which, by expanding the definition of $\app$ and
using the equational theory of delayed substitutions, can be desugared to:

\[\jrhol{\Gamma}{\Psi}{\latern[\xi_1]{(g_1\ t_1\ u_1)}}{\later \Str{\nat}}
          {\latern[\xi_2]{(g_2\ t_2\ u_2)}}{\later \Str{\nat}}
          {\latern[\xi_1, \xi_2, [\res\ltag\leftarrow\res\ltag, \res\rtag \leftarrow \res\rtag]]{(\res\ltag = \res\rtag)}}\]
where, for $i=1,2$:
\[ \xi_i = [g_i \leftarrow \operatorname{zipWith}_i, t_i\leftarrow(tl\ xs_i), u_i \leftarrow (tl\ ys_i)]\]

We apply the \rname{Next} rule, and we have the four following premises:
\begin{itemize}
  \item $\jrhol{\Gamma}{\Psi}{\operatorname{zipWith}_1}{\later A}
          {\operatorname{zipWith}_2}{\later A}
          {\later[\res\ltag\leftarrow\res\ltag, \res\rtag\leftarrow\res\rtag] (\res\ltag = \res\rtag \wedge \forall x y. \res\ltag x y = \res\ltag y x)}$

  \item $\jrhol{\Gamma}{\Psi}{tl\ xs_1}{\later \Str{\nat}}
          {tl\ xs_2}{\later \Str{\nat}}
          {\later[\res\ltag\leftarrow\res\ltag, \res\rtag\leftarrow\res\rtag] (\res\ltag = tl\ ys_2 \wedge \res\rtag = tl\ ys_1)}$

  \item $\jrhol{\Gamma}{\Psi}{tl\ ys_1}{\later \Str{\nat}}
          {tl\ ys_2}{\later \Str{\nat}}
          {\later[\res\ltag\leftarrow\res\ltag, \res\rtag\leftarrow\res\rtag] (\res\ltag = tl\ xs_2 \wedge \res\rtag = tl\ xs_1)}$

  \item $\jrhol{\Gamma; g_1,g_2: A; t_1,t_2,u_1,u_2: \Str{\nat}}
               {\Psi, g_1 = g_2 \wedge \forall x y. g_1 x y = g_1 y x, t_1 = tl\ ys_2 \wedge t_2 = tl\ ys_1,
                     \\ u_1 = tl\ xs_2 \wedge u_2 = tl\ xs_1}
               {g_1\ t_1\ u_1}{\Str{\nat}}
               {g_2\ t_2\ u_2}{\Str{\nat}}
               {\res\ltag = \res\rtag}$

\end{itemize}

To prove the first premise we instantiate the inductive hypothesis we got
from \rname{Fix}. To prove the second and the third premises we use the
equalities $xs_1 = ys_2$, $xs_2 = xs_1$. Finally, the fourth premise is a
simple derivation in HOL that follows from the same equalities plus
the refinements of $g_1,g_2,t_1,t_2,u_1,u_2$. This concludes the
proof.


\subsection{Proof of approximation series}

We now continue with another example that, while still being fully synchronous
(i.e., uses only two-sided rules), goes beyond reasoning about equality of
streams, and showcases the flexibility of streams to represent different kinds
of information and structures.

For instance, streams can be used to represent series of numbers. In
this example, we illustrate an instance of a property about series
that can be proven in our system.  Consider the series $x_0, x_1,
\ldots$ for any $p \geq \frac{1}{2}$ and any $a \geq 0$, where $x_0$
is given and:
\[ x_{i+1} = px_i + (1-p) \frac{a}{x_i} \]
It can be easily shown that if $x_0 \geq \sqrt{a}$, then this series
converges \emph{monotonically} from the top to $\sqrt{a}$. In
particular, $\displaystyle\lim_{i \to \infty} x_i = \sqrt{a}$. (For $p
= \frac{1}{2}$, this is the standard Newton-Raphson series for
square-root computation \cite{Scott11})

The interesting relational property is that for smaller $p$, this
series converges faster. Concretely, define $f(p, a, x_0, i)$ as the
$i$th element of the above series (for the given $p$, $a$ and
$x_0$). Then, the relational property to prove is that:
\[ \forall p_1\, p_2\, a\, x_0\, i.\, (\dfrac{1}{2}\leq p_1 \leq p_2 \conj x_0 \geq \sqrt{a}) \Rightarrow |f(p_1,a,x_0,i) - \sqrt{a}| \leq |f(p_2,a,x_0,i)- \sqrt{a}|\]

We outline the proof of this property. First, note that because
convergence is from the top, $|f(p_2,a,x_0,i) - \sqrt{a}| =
f(p_2,a,x_0,i) - \sqrt{a}$. Therefore, the property above is the same
as:
\[ \forall p_1\, p_2\, a\, x_0\, i.\, (\dfrac{1}{2}\leq p_1 \leq p_2\conj x_0 \geq \sqrt{a}) \Rightarrow f(p_2,a,x_0,i) - f(p_1,a,x_0,i) \geq 0\]
This is easy to establish by induction on $i$.

(Note the importance of the assumption $p \geq \frac{1}{2}$: Without
this assumption, convergence is not monotonic, and this relational
property may not hold. If we start with $x_0 \leq \sqrt{a}$ instead of
$x_0 \geq \sqrt{a}$, we need $p \leq \frac{1}{2}$ for convergence to
be monotonic, this time from below.)

Now we see how we can encode and prove this as a relational property
of a pair of streams.  We can define a stream whose elements are the
elements of one of this series:
\begin{align*}
  \operatorname{approx\_sqrt} &: \real \to \real \to \real \to \Str{\real} \\
  \operatorname{approx\_sqrt} &\defeq \fix{f}{\lambda p. \lambda a. \lambda x.
   \cons{x}{(f \app \later p \app \later a \app \later (p*x + (1-p)*a/x))}}
\end{align*}
We prove:
\[\jgrholnoc{\operatorname{approx\_sqrt_1}}{\real\to\real\to\real\to \Str{\real}}{\operatorname{approx\_sqrt_2}}{\real\to\real\to\real\to \Str{\real}}
         {\Phi}
\]
where
\begin{align*}
\Phi &\defeq \forall p_1 p_2. (\dfrac{1}{2}\leq p_1 \leq p_2) \Rightarrow \forall a_1 a_2. 0 \leq a_1 = a_2 \Rightarrow \forall x_1 x_2. (0 \leq x_1 \leq x_2 \wedge a_1\leq x_1*x_1)
          \\ &\Rightarrow \All(\res\ltag\ p_1\ a_1\ x_1, \res\rtag\ p_2\ a_2\ x_2, \lambda n_1 n_2. 0 \leq n_1 \leq n_2 \wedge a_1 \leq n_1 * n_1) 
\end{align*}
and $\All$ is defined axiomatically as follows:
\[\forall s_1,s_2,n_1,n_2. \phi n_1 n_2 \Rightarrow \later[s'_1\ot s_1, s'_2 \ot s_2]{\All(s'_1,s'_2,\lambda x_1 x_2. \phi)} \Rightarrow \All(\cons{n_1}{s_1}, \cons{n_2}{s_2}, \lambda x_1 x_2. \phi)\]

The meaning of the judgement is that, if we have two approximation series for
the square root of $a$ (formally, we write $a = a_1 = a_2$), with
initial guesses $x_1 \leq x_2$, and parameters $1/2 \leq p_1 \leq
p_2$, then, at every position, the first series is going to be closer
to the root than the second one. Note that we have removed the square
roots in the specification by squaring.

Let $A =\defeq \real\to\real\to\real\to \Str{\real}$.  We will show
how to derive the judgment backwards. The proof starts by applying
\rname{Fix} which has the premise (omitting constant contexts):
\[\jrhol{f_1,f_2 : \later A}{\later[\res\ltag, \res\rtag \leftarrow f_1, f_2]\Phi}{\lambda p_1. \lambda a_1. \lambda x_1. \dots}{A}{\lambda p_2. \lambda a_2. \lambda x_2. \dots}{A}
         {\Phi}
\]
and after applying \rname{Abs} three times:
\[\begin{array}{c}
    \jrhol{f_1,f_2 : \later A; p_1,p_2,a_1,a_2,x_1,x_2 : \real}
        {\\ \later[\res\ltag, \res\rtag \leftarrow f_1, f_2]\Phi, (\dfrac{1}{2}\leq p_1 \leq p_2) , 0 \leq a_1 = a_2 , 0 \leq x_1 \leq x_2, a_1 \leq x_1*x_1}
        {\\ (\lambda y_1.\cons{x_1}{(f_1\app \later p_1\app \later a_1\app \later y_1)})(p_1*x_1 + (1-p_1)*a_1/x_1)}{\Str{\real}}
        {\\ (\lambda y_2.\cons{x_2}{(f_2\app \later p_2\app \later a_2\app \later y_2)})(p_2*x_2 + (1-p_2)*a_2/x_2)}{\Str{\real}}
        {\\ All(\res\ltag, \res\rtag, \lambda n_1 n_2. n_1 \leq n_2)}
\end{array}
\]

Let $\Gamma$ and $\Psi$ denote the typing and logical contexts in the previous judgement. Now we apply \rname{App}, which has two premises:
\begin{itemize}
  \item $\jrhol{\Gamma}{\Psi}
        {\lambda y_1.\cons{x_1}{(f_1\app \later p_1\app \later a_1\app \later y_1)}}{\real\to \Str{\real}}
        {\lambda y_2.\cons{x_2}{(f_2\app \later p_2\app\later a_2\app\later y_2)}}{\real\to \Str{\real}}
        {\\ \forall y_1, y_2. (0 \leq y_1 \leq y_2 \wedge a_1 \leq y_1 * y_1) \Rightarrow All(\res\ltag\ y_1, \res\rtag\ y_2, \lambda n_1 n_2. 0\leq n_1\leq n_2 \wedge a_1 \leq n_1 * n_1)}$
        
  \item $\jrhol{\Gamma}{\Psi}
        {p_1*x_1 + (1-p_1)*a_1/x_1}{\Str{\real}}
        {p_2*x_2 + (1-p_2)*a_2/x_2}{\Str{\real}}
        {\\ 0 \leq \res\ltag \leq \res\rtag \wedge a_1 \leq \res\ltag * \res\ltag}$
\end{itemize}

The second premise can be established in Guarded HOL as an arithmetic
property in our theory of reals. To prove the first one, we start by applying the \rname{Abs}
rule, followed by the \rname{Cons} rule, which has three premises:
\begin{enumerate}
  \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)}
        {x_1}{\real}
        {x_2}{\real}
        {0 \leq \res\ltag \leq \res\rtag \wedge a \leq \res\ltag * \res\ltag}$
  \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)}
        {(f_1\app\later p_1\app\later a_1\app\later y_1)}{\later \Str{\real}}
        {\\ (f_2\app\later p_2\app\later a_2\app\later y_2)}{\later \Str{\real}}
        {\later[\res\ltag \ot \res\ltag, \res\rtag\ot\res\rtag]All(\res\ltag, \res\rtag, \lambda n_1 n_2. 0 \leq n_1 \leq n_2 \wedge a \leq n_1 * n_1)}$
  \item $\jhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)}
        {\forall h_1 h_2 t_1 t_2. 0 \leq h_1 \leq h_2 \Rightarrow a_1 \leq h_1*h_1 \Rightarrow\\
          \later[\res\ltag\ot t_1, \res\rtag \ot t_2] All(\res\ltag, \res\rtag, \lambda n_1 n_2. 0\leq n_1 \leq n_2 \wedge a \leq n_1*n_1) \Rightarrow \\
          All(\cons{h_1}{t_1}, \cons{h_2}{t_2}, \lambda n_1 n_2. 0\leq n_1 \leq n_2 \wedge a \leq n_1*n_1))}$

\end{enumerate}

Premise (1) is just the refinement on $x_1,x_2$, while premise (3) is
the axiomatization of $All$. To prove
premise (2) one instantiates the induction hypothesis given by the
\rname{Fix} rule. In order to do so, we first rewrite the two terms we are comparing to their desugared form:
\[\later[f_1' \ot f_1, p_1' \ot \later p_1, a_1' \ot \later a_1, y_1' \ot \later y_1] f_1'\ p_1'\ a_1'\ y_1'\]
and
\[\later[f_2' \ot f_2, p_2' \ot \later p_2, a_2' \ot\later a_2, y_2' \ot\later y_2] f_2'\ p_2'\ a_2'\ y_2'\]
We can also add by \rname{SUB} the same substitutions to the $\later$
in the conclusion, since the substituted variables do not appear in
the formula. Then we can apply the \rname{Next} rule, which has the premises:
\begin{itemize}
  \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)}
        {f_1}{\later A}
        {f_2}{\later A}
        {\later[\res\ltag\ot\res\ltag, \res\rtag\ot\res\ltag]\Phi}$
  \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)}
        {\later p_1}{\real}
        {\later p_2}{\real}
        {\\ \later[\res\ltag\ot\res\ltag, \res\rtag\ot\res\ltag] \dfrac{1}{2}\leq \res\ltag \leq \res\rtag}$
  \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)}
        {\later a_1}{\real}
        {\later a_2}{\real}
        {\later[\res\ltag\ot\res\ltag, \res\rtag\ot\res\ltag] 0 \leq \res\ltag = \res\rtag}$
  \item $\jrhol{\Gamma, y_1,y_2 : \real}{\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1)}
        {\later y_1}{\real}
        {\later y_2}{\real}
        {\\ \later[\res\ltag\ot\res\ltag, \res\rtag\ot\res\ltag] 0 \leq \res\ltag \leq \res\rtag \wedge a_1' \leq \res\ltag * \res\ltag}$
  \item $\jrhol{\Gamma, y_1,y_2,p_1',p_2',a_1',a_2',y_1',y_2' : \real; f_1', f_2' : A}
        {\Psi, (y_1 \leq y_2 \wedge y_1*y_1 \geq a_1), \Phi\defsubst{f_1'}{f_2'},\\  \dfrac{1}{2}\leq p_1' \leq p_2', 0 \leq a_1' = a_2', 0 \leq y_1' \leq y_2' \wedge a_1' y_1' * y_1'}
        {\\ f_1'\ p_1'\ a_1'\ y_1'}{\Str{\real}}
        {f_2'\ p_2'\ a_2'\ y_2'}{\Str{\real}}{All(\res\ltag, \res\rtag, \lambda n_1 n_2. 0 \leq n_1 \leq n_2 \wedge a \leq n_1 * n_1)}$
 
\end{itemize}

The first four can be proven simply by instantiating and then delaying
one of the axioms. The last one is proven by applying \rname{App} three
times. This concludes the proof.



\subsection{Proof of Cassini's identity}


We continue building on the idea from the previous example of using
streams to represent series of numbers. This time, we prove a
classical identity of the Fibonacci sequence. Since the example
requires to observe the stream at different times, we will also have
to deal with some asynchronicity on the delayed substitutions.

Let $F_n$ be the $n$th Fibonacci number. Cassini's identity states
that $F_{n-1} \cdot F_{n+1} - F_{n}^2 = (-1)^{n}$. Cassini's identity
can be stated as a stream problem as follows. First, let $F$ be the
Fibonnaci stream ($1,1,2,3,5,\ldots$) and $A$ be the stream
$1,-1,1,-1,\ldots$ Let $\oplus$ and $\otimes$ be infix functions that
add and multiply two streams pointwise. Cassini's identity can then be
informally written as:
\[ F \otimes  \tl(\tl\; F) = \tl(F \otimes F) \oplus A \]

In order to formalize Cassini's identity in our system, we first define:
\begin{align*}
  \oplus &: \Str{\nat} \to \Str{\nat} \to \Str{\nat}   &\otimes &: \Str{\nat} \to \Str{\nat} \to \Str{\nat} \\                            
  \oplus &\defeq \begin{array}{l}\fix{f}{\lambda s.\lambda t\\ \cons{(\hd\ x + \hd\ y)}{(f \app (\tl\ x) \app (\tl\ y))}}\end{array} \quad \quad
  &\otimes &\defeq \begin{array}{l}\fix{f}{\lambda s.\lambda t\\ \cons{(\hd\ x * \hd\ y)}{(f \app (\tl\ x) \app (\tl\ y))}}\end{array}
\end{align*}
Then we define $F$ and $A$ as the fixpoints of the equations:
\begin{align*}
  F &\defeq \fix{F}{\cons{1}{\later[ F' \ot F] (\cons{1}{\later [ T \ot \tl\ F' ] ( F' \oplus T)})}} \\
  A &\defeq \fix{A}{\cons{1}{\later(\cons{-1}{A})}}
\end{align*}


We prove (using prefix notation for $\oplus$ and $\otimes$):
\[\jgrholnoc
           {\later [ T_1 \ot \tl\ F] \otimes \app (\later F) \app \tl\ T_1 }{\later\later \Str{\nat}}
           {\oplus \app \tl(F \otimes F) \app (\later A)}{\later \Str{\nat}}
           { \res\ltag = \later \res\rtag}
\]


The proof combines applications of two-sided rules and one-sided
rules; in particular, we use the rule \rname{NEXT-L} to proceed with
the proof for a judgement where the left expression is delayed twice
and the right expression is delayed once.

By conversion, in the logic we can prove the following equalities:
\[
  \Psi \defeq \left\{\begin{aligned}
  F &= \cons{1}{\later(\cons{1}{\later [ T \ot tl\ F ] ( F \oplus T)})}, \\
  A &= \cons{1}{\later(\cons{-1}{\later A})}
  \end{aligned}\right\}
\]

Using these equalities, and desugaring the applications, the judgment we want to prove is (omitting constant contexts):

\[\begin{array}{c}\jrhol{F,A: \Str{\nat}}{\Psi}
           {\later [ T_1 \ot tl\ F] \later [T_1' \ot tl\ T_1] (F \otimes T_1') }{\later\later \Str{\nat}}
           {\\ \later [T_2 \ot tl(F \otimes F)] (T_2 \oplus A)}{\later \Str{\nat}}
           {\\ \later[\res\ltag' \ot \res\ltag, T_1 \ot tl\ F] \later [\res\ltag''\ot \res\ltag, \res\rtag'\ot\res\rtag, T_1' \ot tl\ T_1, T_2' \ot tl(F \otimes F)] \res\ltag'' = \res\rtag'}
\end{array}
\]

Notice that on the left, since we want to apply tail twice to $F$, we
need to delay the term twice so that $F$ and $tl\ tl\ F$ have the same
type. On the right, we just need to delay the term once. As for the
logical conclusion, $\res\ltag$ needs to be delayed twice, while
$\res\rtag$ only once. The way to do this is by having $\res\ltag$
appear on the two substitutions but $\res\rtag$ only on the inner one.

We start by applying \rname{NEXT-L}, which has the two following premises:
\begin{itemize}
  \item $\juhol{F,A: \Str{\nat}}{\Psi}{tl\ F }{\later \Str{\nat}}{\later[\res' \ot \res] tl\ F = \later \res'}$

  \item $\jrhol{F,A, T_1: \Str{\nat}}{\Psi, tl\ F = \later T_1}{\later [T_1' \ot tl\ T_1] (F \otimes T_1') }{\later \Str{\nat}}
              {\\ \later [T_2 \ot tl(F \otimes F)] (T_2 \oplus A)}{\later \Str{\nat}}
              {\later [\res\ltag''\ot \res\ltag, \res\rtag'\ot\res\rtag, T_1' \ot tl\ T_1, T_2' \ot tl(F \otimes F)] \res\ltag'' = \res\rtag'}$

\end{itemize}

The first premise is trivial. We continue by applying \rname{NEXT} to the second, which has the following premises:

\begin{itemize}
  \item $\jrhol{F,A, T_1: \Str{\nat}}{\Psi, tl\ F = \later T_1}{tl\ T_1}{\later \Str{\nat}}{tl(F \otimes F)}{\later \Str{\nat}}
        {\\ \later[\res\ltag' \ot \res\ltag, \res\rtag' \ot \res\rtag] T_1 = \later \res\ltag' \wedge \res\ltag'\otimes\res\ltag' = \res\rtag'}$

  \item $\jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2}{F \otimes T_1'}{\Str{\nat}}{T_2 \oplus A}{\Str{\nat}}{\res\ltag = \res\rtag}$

\end{itemize}

Again, the first premise is trivial. We apply \rname{APP} twice to the second, and we have to prove:

\begin{itemize}
  \item $\jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2}{F}{\Str{\nat}}{A}{\Str{\nat}}{\\ \res\ltag = F \wedge \res\rtag = A}$

  \item $\jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2}{T_1'}{\Str{\nat}}{T_2}{\Str{\nat}}
        {\\ F = \cons{1}{\later(\cons{1}{\later T_1})} \wedge \res\ltag \otimes \res\rtag = \res\rtag}$

  \item $\jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2}
         {\\ \otimes}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}{\oplus}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}
         {\forall X_1 X_2 Y_1 Y_2. X_1 = F \wedge X_2 = A \Rightarrow F = \cons{1}{\later(\cons{1}{\later Y_1})} \wedge Y_1 \otimes Y_1 = Y_2 \Rightarrow \res\ltag\ X_1\ Y_1 = \res\rtag\ X_2\ Y_2}$
\end{itemize}

The two first premises are easy to prove. We will show how to prove the last one. 
For this, we need a stronger induction hypothesis for $\hat{\oplus}$ and
$\hat{\otimes}$. We propose the following:
\[\begin{array}{c}
  \forall g_1,g_2,b_1,G,B. G = \conshat{g_1}{\conshat{g_2}{(G\hat{\oplus} (\hat{tl} G))}} \wedge b_1 = g_1^2 + g_1 g_2 - g_2^2 \wedge B = \conshat{b_1}{\conshat{-b_1}{B}}
   \\  \Rightarrow G \hat{\otimes} \hat{tl}(\hat{tl} G) = \hat{tl} (G \hat{\otimes} G) \hat{\oplus} B 
\end{array}
\]

We then use the \rname{SUB} rule to strengthen the inductive hypothesis, and now the new judgement to prove is:

\[\begin{array}{c}
  \jrhol{F,A, T_1', T_2: \Str{\nat}}{\Psi,tl\ F = \later T_1 , tl\ T_1 = \later T_1', T_1'\otimes T_1' = T_2}
         {\\ \otimes}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}{\oplus}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}
         {\\ \forall X_1 X_2 Y_1 Y_2. (\exists g_1, g_2, b_1, G, B. G = \cons{g_1}{\later(\cons{g_2}{\later [ G' \ot tl\ G ] ( G \oplus G')})}
           \wedge b_1 = g_1^2 + g_1 g_2 - g_2^2 \wedge \\ B = \cons{b_1}{\later(\cons{-b_1}{\later B})} \wedge
          X_1 = G \wedge X_2 = B \wedge X_1 = \cons{1}{\later(\cons{1}{\later Y_1})} \wedge Y_1 \otimes Y_1 = Y_2) \Rightarrow \res\ltag\ X_1\ Y_1 = \res\rtag\ X_2\ Y_2}
\end{array}
\]

Let $\Gamma'$, $\Psi'$ and $\Phi_{IH}$ denote respectively the typing context, logical context and logical conclusion of the previous judgement. The premise
of the FIX rule is:

\[\begin{array}{c}
  \jrhol{\Gamma; f_1,f_2 : \later(\Str{\nat} \to \Str{\nat} \to \Str{\nat})}{\Psi',\later[\res\ltag \ot f_1, \res\rtag \ot f_2] \Phi_{IH}}
         {\\ \fix{f_1}{\lambda X_1. \lambda Y_1. \dots}}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}{\fix{f_2}{\lambda X_2. \lambda Y_2. \dots}}{\Str{\nat} \to \Str{\nat} \to \Str{\nat}}{\Phi_{IH}}
\end{array}
\]

Let $\Phi_E$ denote the existential clause in $\Phi_{IH}$. After applying \rname{ABS} twice, we have:

\[\begin{array}{c}
   \jrhol{\Gamma; f_1,f_2 : \later(\Str{\nat} \to \Str{\nat} \to \Str{\nat}); X_1,X_2,Y_1,Y_2 : \Str{\nat}}{\Psi',\later[\res\ltag \ot f_1, \res\rtag \ot f_2] \Phi_{IH}, \Phi_E}
         {\\ \cons{(hd\ X_1)*(hd\ Y_1)}{f_1 \app (tl\ X_1) \app (tl\ Y_1)}}{\Str{\nat}}{\cons{(hd\ X_2)+(hd\ Y_2)}{f_2 \app (tl\ X_2) \app (tl\ Y_2)}}{\Str{\nat}}{\res\ltag = \res\rtag}
\end{array}\]

And then we apply \rname{Cons} to prove equality on the heads and the tails:

\begin{itemize}
\item  $\jrhol{\Gamma; f_1,f_2 : \later(\Str{\nat} \to \Str{\nat} \to \Str{\nat}); X_1,X_2,Y_1,Y_2 : \Str{\nat}}{\Psi',\later[\res\ltag \ot f_1, \res\rtag \ot f_2] \Phi_{IH}, \Phi_E}
         {\\ (hd\ X_1)*(hd\ Y_1)}{\nat}{(hd\ X_2)+(hd\ Y_2)}{\nat}{\res\ltag = \res\rtag}$
\item  $\jrhol{\Gamma; f_1,f_2 : \later(\Str{\nat} \to \Str{\nat} \to \Str{\nat}); X_1,X_2,Y_1,Y_2 : \Str{\nat}}{\Psi',\later[\res\ltag \ot f_1, \res\rtag \ot f_2] \Phi_{IH}, \Phi_E}
         {\\ f_1 \app (tl\ X_1) \app (tl\ Y_1)}{\later \Str{\nat}}{f_2 \app (tl\ X_2) \app (tl\ Y_2)}{\later \Str{\nat}}{\res\ltag = \res\rtag}$
\end{itemize}

To prove the first one we notice that $hd X_1 * hd Y_1 = g_1 * (g_1 + g_2) = g_1^2 + g_2*g_1 =  g_2^2 + g_1^2 + g_1 * g_2 - g_2^2 = hd X_2 * hd Y_2$. To prove the second one
we need to check that $tl X_1,tl Y_1, tl X_2,tl Y_2$ satisfy the precondition of the inductive hypothesis. In particular, we need to check that
\[-b_1 = -g_1^2 - g_1 g_2 + g_2^2 = g_2^2 + g_2 (g_1 + g_2) - (g_1 + g_2)^2\]
which is can be proven by arithmetic computation.

 

\section{Unary fragment}

In this section we introduce a unary system to prove properties
about a single term of the guarded lambda calculus. We will start
by adding some definitions Guarded HOL for the unary diamond monad,
following by the derivation rules for both the non-probabilistic
and the probabilistic system, plus the metatheory and an example.

\subsection{Unary fragment of GHOL}

The unary semantics of the diamond monad are:

\[\sem{\diamond_{[x\leftarrow t]}\phi}_i\defeq \left\{(\delta,\gamma) \isetsep
    \operatorname{Pr}_{v\ot \left(\sem{t}_i(\delta,\gamma)\right)} [(\delta, (\gamma, v))\in \sem{\phi}_i] = 1 \right\} \]

The rules are on Figure~\ref{fig:u-prob-ghol}
    
\begin{figure*}[!htb]
\small
\begin{mathpar}
  \infer[\sf MONO1]
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x\gets t]}\psi}}
  {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x\gets t]}\phi} \and
    \jghol{\Delta}{\Sigma}{\Gamma, x : C}{\Psi, \phi}{\psi}
  }

\and

  \infer[\sf UNIT1]{\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{
    \diamond_{[x\leftarrow \munit{t}]} \phi}}
       {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{
           \phi[t/x]}}
\and
 
\infer[\sf MLET1]
      {\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{
          \diamond_{[y\leftarrow \mlet{x}{t}{t'}]} \psi}}{
          \jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\diamond_{[x\leftarrow t]} \phi}
          \and
          \jghol{\Delta}{\Sigma}{\Gamma,x:C}{\Psi,\phi}{
            \diamond_{[y\leftarrow t']}\psi}}


\end{mathpar}
\caption{Rules for the unary diamond modality}\label{fig:u-prob-ghol}
\end{figure*}
        




\subsection{Guarded UHOL}

We start by defining the Guarded UHOL system, which allows us to prove
logical properties of a term of the Guarded Lambda Calculus.  More
concretely, judgements have the form:
\[ \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\sigma}{\phi} \]
where $t$ is a term well-typed in the dual context $\Delta \mid \Gamma$ and $\phi$ is a logical formula well-typed in the context
$\Delta \mid \Gamma, \res : \sigma$ and that can refer to $t$ via the special variable $\res$. The logical contexts $\Sigma$ and
$\Psi$ consist respectively of refinements over the contexts $\Delta$ and $\Gamma$.


\subsection{Derivation rules}

The rule \rname{Next} corresponds to the introduction of the later
modality. A refinement $\Phi_i$ is proven on every term in the
substitution, and using those as a premise, a refinement $\Phi$ is
proven on $t$. In the notation $\later[\res \ot \res]\Phi$ the first
$\res$ is the variable bound by the delayed substitution inside $\Phi$
while the second $\res$ is the distinguished variable in the refinement
that refers to the term that is being typed. In other words, $t$ satisfies
$\later[\res \ot \res]\Phi$ if $\later[\res \ot t]\Phi$.
The rule \rname{Prev} corresponds to the elimination of the later modality.
If we can prove $\later \phi$ in a constant context, then we can also prove
$\phi$.
The rule \rname{Box} applies the constant modality on a formula that can
be proven on a constant context.
The rule \rname{LetBox} removes the constant modality from a formula $\Phi$
by using it as a constant premise to prove another formula $\Phi'$.
The rule \rname{LetConst} shifts constant terms between contexts.
The rule \rname{Fix} introduces a fixpoint and proves a refinement on it by
Loeb induction.
The rule \rname{Cons} proves a property on a stream from a refinement on its
head and its tail.
The rule \rname{ConsHat} is the analogue of \rname{Cons} to build constant streams.
In particular, the $\hat{::}$ operator can be defined as $\lambda x. \lambda s. \letbox{(y,t)}{(x,s)}{\boxx(\cons{y}{\later{t}})}$.
Conversely the rules \rname{Head} and \rname{Tail} respectively prove a property
on the head and the tail of a stream from a property on the full stream.


\begin{figure*}[!htb]
\small
\begin{mathpar}
\inferrule*[Right=\sf Next]
      {\jguhol{\Delta}{\Sigma}{\Gamma, x_1:A_1, \dots, x_n:A_n}{\Psi, \Phi_1\subst{\res}{x_1}, \dots, \Phi_n\subst{\res}{x_n}}{t}{A}{\Phi} \\
       \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\later{A_1}}{\dsubst{\res}{\res}{\Phi_1}}
       \\ \dots \\
       \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_n}{\later{A_n}}{\dsubst{\res}{\res}{\Phi_n}}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\nextt{x_1 \leftarrow t_1, \dots, x_n \leftarrow t_n}{t}}{\later{A}}{\dsubst{x_1,\dots,x_n,\res}{t_1,\dots,t_n,\res}{\Phi}}}
      


 \inferrule*[Right=\sf Prev]
      {\jguhol{\Delta}{\Sigma}{\cdot}{\cdot}{t}{\later A}{\dsubst{\res}{\res}{\Phi}}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\prev{t}}{A}{\Phi}} 

 \inferrule*[Right=\sf Box]
      {\jguhol{\Delta}{\Sigma}{\cdot}{\cdot}{t}{A}{\Phi}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\boxx{t}}{\square A}{\square \Phi\subst{\res}{\letbox{x}{\res}{x}}}} 

 \inferrule*[Right=\sf LetBox]
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u}{\square B}{\square \Phi\subst{\res}{\letbox{x}{\res}{x}}} \\
       \jguhol{\Delta, x : B}{\Sigma, \Phi\subst{\res}{x}}{\Gamma}{\Psi}{t}{A}{\Phi'}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\letbox{x}{u}{t}}{A}{\Phi'}} 

     
  \inferrule*[Right=\sf LetConst]
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u}{B}{\Phi} \\
       \jguhol{\Delta, x : B}{\Sigma, \Phi\subst{\res}{x}}{\Gamma}{\Psi}{t}{A}{\Phi'} \\
       B,\Phi\ \text{constant} \\ FV(\Phi)\cap FV(\Gamma) = \emptyset}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\letconst{x}{u}{t}}{A}{\Phi'}} 

      
  \inferrule*[Right=\sf Fix]
      {\jguhol{\Delta}{\Sigma}{\Gamma, f:\later{A}}{\dsubst{\res}{f}{\Phi}}{t}{A}{\Phi}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{fix f. t}{A}{\Phi}} 

  \inferrule*[Right=\sf Cons]
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x}{A}{\Phi_h} \\
       \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{xs}{\later{\Str{A}}}{\Phi_t} \\
       \jhol{\Gamma}{\Psi}{\forall x,xs. \Phi_h\subst{\res}{x} \Rightarrow \Phi_t\subst{\res}{xs} \Rightarrow \Phi\subst{\res}{\cons{x}{xs}}}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\cons{x}{xs}}{\Str{A}}{\Phi}}

  \inferrule*[Right=\sf ConsHat]
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{x}{A}{\Phi_h} \\
       \jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{xs}{\square \Str{A}}{\square \Phi_t} \\
       \jhol{\Gamma}{\Psi}{\forall x,xs. \Phi_h\subst{\res}{x} \Rightarrow \Phi_t\subst{\res}{xs} \Rightarrow \Phi\subst{\res}{\conshat{x}{xs}}} \\
      A,\Phi_h \text{ constant} }
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\conshat{x}{xs}}{\square \Str{A}}{\square \Phi}} 


  \inferrule*[Right=\sf Head]
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\Str{A}}{\Phi\subst{\res}{hd\ \res}}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{hd\ t}{A}{\Phi}} 
      
   \inferrule*[Right=\sf Tail ]
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\Str{A}}{\Phi\subst{\res}{tl\ \res}}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{tl\ t}{\later{Str_A}}{\Phi}} 

      
\end{mathpar}
\caption{Guarded Unary Higher-Order Logic rules}\label{fig:uhol}
\end{figure*}


The intended meaning for a judgment $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\tau}{\phi}$ is:
``For every valuations $\delta$, $\gamma$ of $\Delta$ and $\Gamma$,
\[\sem{\Delta\mid\Gamma\vdash \square\Sigma}(\delta,\gamma) \wedge \sem{\Delta\mid\Gamma\vdash \Psi}(\delta,\gamma) \To
 \sem{\Delta\mid\Gamma, \res:\tau\vdash \Sigma}(\delta,\langle \gamma, \sem{\Delta\mid\Gamma\vdash t}(\delta,\gamma)\rangle) \text{''} \]


\subsection{Metatheory}

We now the most interesting metatheoretical properties of Guarded UHOL. In particular, Guarded UHOL
is equivalent to Guarded HOL:

\begin{theorem}[Equivalence with Guarded HOL] \label{thm:equiv-uhol-hol}
  For every contexts $\Delta,\Gamma$, type $\sigma$, term $t$, sets of assertions $\Sigma,\Psi$
  and assertion $\phi$, the following are equivalent:
  \begin{itemize}
  \item
    $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\sigma}{\phi}$
  \item
    $\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{\phi\subst{\res}{t}}$
  \end{itemize}
\end{theorem}
The proof is analogous to the relational case

The previous result allows us to lift the soundness result from Guarded HOL to
Guarded UHOL.

\begin{corollary}[Soundness and consistency]\label{cor:uhol:sound}
If $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\sigma}{\phi}$, then for every valuations
$\delta \models \Delta$, $\gamma\models\Gamma$:
\[\sem{\Delta \vdash \Sigma}(\delta) \wedge \sem{\Delta \mid \Gamma \vdash \Psi}(\delta,\gamma) \Rightarrow
    \sem{\Delta \mid \Gamma,\res:\sigma \vdash \phi}(\delta, \gamma[\res \ot \sem{\Delta \mid \Gamma \vdash t}(\delta,\gamma)])\]
In particular, there is no proof of
$\jguhol{\Delta}{\emptyset}{\Gamma}{\emptyset}{t}{\sigma}{\bot}$ in Guarded UHOL.
\end{corollary}








\subsection{Probabilistic extension}

We comment on the rules, starting from the rules of the unary
logic. There are three new rules for the probabilistic case, and they
all establish that an expression $u$ of type $\Distr(D)$ satisfies the
assertion $\diamond_{[y\leftarrow \res]} \phi$, i.e.\, for every
element $v$ in the support of (the interpretation of) $u$, the
interpretation of $\phi$ with the valuation $[y \mapsto v]$ is
true. This intuition is captured by the rule $[\textsf{SUPP}]$, which
can be used in particular in case $u$ is a primitive distribution. The
rule [\textsf{UNIT}] considers the case where $u$ is of the form
$\munit{t}$; in this case, it is clearly sufficient to know that
$\phi[t/y]$ is valid. The rule [\textsf{MLET}] simply captures the
fact that the support of $\mlet{x}{t}{t'}$ is the disjoint union of
the support of $t'$ under all the assignments of $x$ to values in the
support of $t$.


\begin{figure*}[!htb]
\small

\textbf{Guarded UHOL}
\begin{mathpar}
\infer[\sf UNIT]
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\munit{t}}{\Distr(C)}{
          \diamond_{[y\leftarrow \res]} \Phi}}
      {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{C}{\Phi
          [\res/y]}}
\and    
\infer[\sf MLET]
    {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{\mlet{x}{t}{t'}}{
        \Distr(D)}{\diamond_{[y\leftarrow \res]} \psi}}
    {\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t}{\Distr(C)}{\diamond_{[x\leftarrow r]} \Phi}
     \and
     \jguhol{\Delta}{\Sigma}{\Gamma,x:C}{\Psi,\phi
     }{t'}{\Distr(D)}{\diamond_{[y\leftarrow \res]} \psi}}



 \infer[\sf SUPP]{\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{u}{\Distr(D)}{
     \diamond_{[y\leftarrow \res]}\phi}}{\jghol{\Delta}{\Sigma}{\Gamma}{\Psi}{
     \Pr_{z\sim u}[\phi[z/y]]=1}}
 
\end{mathpar}

\caption{Proof rules for probabilistic constructs -- unary case}\label{fig:puhol}
\end{figure*}


Finally, we prove an embedding lemma for Guarded UHOL. The proof can be
carried by induction on the structure of derivations, or using the
equivalence between Guarded UHOL and Guarded HOL (Theorem~\ref{thm:equiv-uhol-hol}).
\begin{lem}[Embedding lemma]\label{lem:emb-uhol-rhol} Assume that:
\begin{itemize}
\item $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{\phi}$
\item $\jguhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_2}{\sigma_2}{\phi'}$
\end{itemize}
Then
$\jgrhol{\Delta}{\Sigma}{\Gamma}{\Psi}{t_1}{\sigma_1}{t_2}{\sigma_2}{
  \phi\subst{\res}{\res\ltag}\land \phi'\subst{\res}{\res\rtag}}$.
\end{lem}


\subsection{Unary example: Every two}

We define the $every2$ function, which receives a stream and returns another stream consisting
of the elements at even positions in the input stream. Note that this function, while productive,
cannot be built with the type $Str \to Str$, since we need to take twice the tail of the argument,
which would have type $\later\later Str$, and then a $Str$ cannot be built. Instead, we need to use the constant modality
as follows:

\begin{align*}
every2 &: \square Str \to Str \\
every2 &\defeq \fix{every2}{\lambda s. \cons{\hat{hd}(\hat{tl}\ s)}{(every2 \app {\rm next} (\hat{tl}(\hat{tl}\ s)))}}
\end{align*}
Where the $\hat{hd}$ and $\hat{tl}$ functions are not the native ones, but rather they are defined as:
\begin{align*}
  \hat{hd} &: \square Str \to \nat         &\hat{tl} &: \square Str \to \square Str \\              
  \hat{hd} &\defeq \lambda s. \letbox{x}{s}{hd\ x} \quad\quad     &\hat{tl} &\defeq \lambda s. \letbox{x}{s}{\boxx{(\prev{(tl\ x)})}}
\end{align*}
The property we want to prove is:
\[
\jguhol{\cdot}{\cdot}{ones : \square Str}{\Psi}{every2}{\square Str \to Str}{\forall s. s = ones \Rightarrow \res\ s = (\letbox{x}{s}{x})}
\]
where $ones$ is the constant stream containing only the number 1 defined as:
\[
ones \defeq \boxx{(\fix{f}{\cons{1}{f}})}
\]
For which we can prove the following properties:
\[
\Psi \defeq \hat{hd}\ ones = 1, \hat{tl}\ ones = ones
\]
In the rest of the proof we omit the empty contexts $\Delta$ and $\Sigma$.
We start by applying the \rname{Fix} rule, which has the premise:
\[
\begin{array}{c}
\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}}{\Psi, \later[\res \ot every2]{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}}
                     {\\ \lambda s. (\cdots)}{\square Str \to Str}{\forall s. s = ones \Rightarrow (\res\ s) = \letbox{x}{s}{x}}
\end{array}
\]
We apply the \rname{Abs} rule inmediately after:
\[
\begin{array}{c}
\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str}
       {\\ \Psi, \later[\res \ot every2]{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones}
                     {\\ \cons{\hat{hd}(\hat{tl}\ s)}{(every2 \app \later (\hat{tl}(\hat{tl}\ s)))}}{Str}{\res = \letbox{x}{s}{x}}
\end{array}
\]

By \rname{SUB} and the equivalence $\letbox{x}{s}{x} \equiv \letbox{x}{ones}{x}$, we can change the
conclusion of the judgement.
Now we use the \rname{Cons} rule, which has three premises:
\begin{enumerate}
  \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str}
               {\\ \Psi, \later[\res \ot every2]{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones}
                             { \hat{hd}(\hat{tl}\ s)}{\nat}{\res = 1}$
\\
  \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str}
               {\\ \Psi, \later[\res \ot every2]{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones}
                             {\\(every2 \app \later(\hat{tl}(\hat{tl}\ s)))}{Str}{\dsubst{\res}{\res}{\res = \letbox{x}{ones}{x}}}$
\\
  \item $\jhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str}
             {\\ \Psi, \dsubst{\res}{every2}{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones}
                             {\\ \forall y, ys. y = 1 \Rightarrow \later[zs \ot ys]{(zs = \letbox{x}{ones}{x})} \Rightarrow \cons{y}{ys} = (\letbox{x}{ones}{x})}$
    
\end{enumerate}

Premises (1) is a consequence of the properties of $ones$. To prove premise (3)
we reduce the letbox with the box inside $ones$, and do some reasoning using the definition of the fixpoint. To prove the
premise (2) we first desugar the term we are typing:
\[every2 \app \later(\hat{tl}(\hat{tl}\ s))) \defeq \later[g \leftarrow every2, t \leftarrow \later(\hat{tl}(\hat{tl}\ s))]{g t} \]
and then we apply \rname{Next} which has the following premises:
\begin{itemize}
  \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str}
               {\\ \Psi, \dsubst{\res}{every2}{\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x}}, s = ones}
                             {\\ every2}{\later{(\square Str \to Str)}}{\dsubst{\res}{\res}{(\forall s = ones \Rightarrow \res\ s = \letbox{x}{s}{x})}}$
\\
  \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str}
               {\\ \Psi, \dsubst{\res}{every2}{(\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x})}, s = ones}
                             {\\ \later(\hat{tl}(\hat{tl}\ s))}{\later\square Str}{\dsubst{\res}{\res}{(\res = ones)}}$
\\
  \item $\juhol{ones : \square Str, every2:\later{(\square Str \to Str)}, s: \square Str, g: \square Str \to Str, t: \square Str}
              {\\ \Psi, \dsubst{\res}{every2}{(\forall s. s = ones \Rightarrow \res\ s = \letbox{x}{s}{x})}, s = ones,
                \\ \forall s. s = ones \Rightarrow g\ s = \letbox{x}{s}{x}, t = ones}
                             {g\ t}{Str}{\res = (\letbox{x}{ones}{x})}$
    
\end{itemize}
The first premise is just an application of the \rname{Var} rule.  The
second premise can be proven as a consequence of the properties of
$ones$.  Finally, the third premise can be proven with some simple
logical reasoning in HOL.  This concludes the proof.

 


     
 

\end{document}