File size: 106,654 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
1889
1890
1891
1892
1893
1894
1895
1896
1897
1898
1899
1900
1901
1902
1903
1904
1905
1906
1907
1908
1909
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919
1920
1921
1922
1923
1924
1925
1926
1927
1928
1929
1930
1931
1932
1933
1934
1935
1936
1937
1938
1939
1940
1941
1942
1943
1944
1945
1946
1947
1948
1949
1950
1951
1952
1953
1954
1955
1956
1957
1958
1959
1960
1961
1962
1963
1964
1965
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
2026
2027
{
    "paper_id": "2022",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T07:32:53.796546Z"
    },
    "title": "Data-to-text systems as writing environment",
    "authors": [
        {
            "first": "Adela",
            "middle": [],
            "last": "Schneider",
            "suffix": "",
            "affiliation": {},
            "email": ""
        },
        {
            "first": "Andreas",
            "middle": [],
            "last": "Madsack",
            "suffix": "",
            "affiliation": {},
            "email": ""
        },
        {
            "first": "Johanna",
            "middle": [],
            "last": "Heininger",
            "suffix": "",
            "affiliation": {},
            "email": ""
        },
        {
            "first": "Ching-Yi",
            "middle": [],
            "last": "Chen",
            "suffix": "",
            "affiliation": {},
            "email": ""
        },
        {
            "first": "Robert",
            "middle": [],
            "last": "Wei\u00dfgraeber",
            "suffix": "",
            "affiliation": {},
            "email": ""
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "Today, data-to-text systems are used as commercial solutions for automated text production of large quantities of text. Therefore, they already represent a new technology of writing. This new technology requires the author, as an act of writing, both to configure a system that then takes over the transformation into a real text, but also to maintain strategies of traditional writing. What should an environment look like, where a human guides a machine to write texts? Based on a comparison of the NLG pipeline architecture with the results of the research on the human writing process, this paper attempts to take an overview of which tasks need to be solved and which strategies are necessary to produce good texts in this environment. From this synopsis, principles for the design of data-to-text systems as a functioning writing environment are then derived.",
    "pdf_parse": {
        "paper_id": "2022",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "Today, data-to-text systems are used as commercial solutions for automated text production of large quantities of text. Therefore, they already represent a new technology of writing. This new technology requires the author, as an act of writing, both to configure a system that then takes over the transformation into a real text, but also to maintain strategies of traditional writing. What should an environment look like, where a human guides a machine to write texts? Based on a comparison of the NLG pipeline architecture with the results of the research on the human writing process, this paper attempts to take an overview of which tasks need to be solved and which strategies are necessary to produce good texts in this environment. From this synopsis, principles for the design of data-to-text systems as a functioning writing environment are then derived.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "Natural Language Generation (NLG) systems are computer systems that automatically generate texts in human languages, using advanced techniques from artificial intelligence and/or computational linguistics (Carlson, 2015) . Non-academic NLG systems are used in different areas of text production and result in fundamental changes for content creation and publication processes: They form a new type of writing technology and create a new environment for humans in which texts are generated automatically, but humans still (co-)design the rules and specifications for this generation.",
                "cite_spans": [
                    {
                        "start": 205,
                        "end": 220,
                        "text": "(Carlson, 2015)",
                        "ref_id": "BIBREF1"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "While NLG systems based on pre-trained large language models function more as writing assistants for authors on an individual level, the NLG systems that are the subject of this study have a different aim: They are configured to be able to produce large amounts of text automatically.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "In this context, writing is regarded in a broader sense and means creating a blueprint for producing specific texts. So this new type of writing can be described as meta-writing: However, since the requirements of text structure, expression, and realisation of a communication goal cannot be solved on an abstract level only, many traditional writing tasks remain to be done by the author. Mahlow and Dale (2014) have described this new condition as follows: \"Automated text production -when the author is not the writer\". This observation raises the question, what a writing environment should be like in which a machine is guided by an author to write a text?",
                "cite_spans": [
                    {
                        "start": 390,
                        "end": 412,
                        "text": "Mahlow and Dale (2014)",
                        "ref_id": "BIBREF15"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "In this research, we use the framework of creating a writing environment to set out the requirements for an NLG system. So, the human writer is considered here as the agent, while the software functions as the environment. This setting is due to the fact that writing, in general, is primarily perceived as an individual action, even though some instances of writing are performed in collaboration. But of course, it is not the only possible framework. The interaction between humans and machines has recently been discussed, especially in the communicative field of AI, where both humans and the instances of AI are seen as agents and the aspect of collaboration is much more prominent (for the field of journalism: Lewis et al. (2019) ; for fiction writing: Manjavacas et al. (2017) ; Clark et al. (2018) ).",
                "cite_spans": [
                    {
                        "start": 717,
                        "end": 736,
                        "text": "Lewis et al. (2019)",
                        "ref_id": "BIBREF14"
                    },
                    {
                        "start": 760,
                        "end": 784,
                        "text": "Manjavacas et al. (2017)",
                        "ref_id": "BIBREF16"
                    },
                    {
                        "start": 787,
                        "end": 806,
                        "text": "Clark et al. (2018)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "And indeed, it may be that statistical approaches and deep-learning methods, in particular, bring the software's autonomy more to the fore. Autonomy is, after all, the distinctive property of the agent (Henrickson, 2018) . This then would call for a reassessment of the situation, looking more closely at the requirements of collaboration within this described environment. However, data-to-text systems in real-world applications still require such a share of human configuration and control and the creative contribution share of the software, at least in the NLG systems focused on in this paper, is still so limited that it would not be adequate to claim 1 creative autonomy of the software in the process.",
                "cite_spans": [
                    {
                        "start": 202,
                        "end": 220,
                        "text": "(Henrickson, 2018)",
                        "ref_id": "BIBREF11"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "The environmental framework with its orientation towards the writing processes also offers the advantage of shifting the focus in the evaluation of NLG systems (Howcroft et al., 2020) from the evaluation of the output to an evaluation of the processes, that Gehrmann et al. (2022) recently postulated: \"Evaluating NLG tasks only through the lens of outputs is thus insufficient and we should strife (sic) to deliver a more fine-grained breakdown (...)\". For traditional writing it is set that the principle of having control over the writing and editorial processes is the most effective method of influencing text quality (Wyss, 2013; Perrin, 2001 ). And we assume it remains valid also for working with NLG systems. Thus, our approach could open up new perspectives for the evaluation of NLG systems.",
                "cite_spans": [
                    {
                        "start": 160,
                        "end": 183,
                        "text": "(Howcroft et al., 2020)",
                        "ref_id": null
                    },
                    {
                        "start": 258,
                        "end": 280,
                        "text": "Gehrmann et al. (2022)",
                        "ref_id": "BIBREF9"
                    },
                    {
                        "start": 623,
                        "end": 635,
                        "text": "(Wyss, 2013;",
                        "ref_id": "BIBREF27"
                    },
                    {
                        "start": 636,
                        "end": 648,
                        "text": "Perrin, 2001",
                        "ref_id": "BIBREF19"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "What Perrin stated in 2002 for writing per se also applies to automated text production nowadays: \"Writing is thus changing from a field of largely intuitive language design to a language technology that becomes aware of its compositional principles and purposefully uses its means, tools, and strategies\" (translated from German (Perrin, 2002, page 7) ).",
                "cite_spans": [
                    {
                        "start": 330,
                        "end": 352,
                        "text": "(Perrin, 2002, page 7)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "As a starting point to achieve such an awareness and methodology for this new kind of writing, including a system of rules, strategies, and cues that guide action, we want to make the action steps, tools and decisions within the processes explicit:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "1. In order to approach this, we take a look at the structure and design of NLG systems, because from these the special requirements and conditions are derived to which the user is subject with their text generation task. (The different categories of NLG systems and Overview of the NLG pipeline)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "2. To identify the factors that are conducive to the production of (good) texts, we will outline how the human writing process is organized (A model of the human writing process). In doing so, we will refer to the results of writing process research as well as to the approaches to the development of modern writing software. (Requirements for writing software)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "3. With these findings in mind, we try to take a closer look at automated text production with NLG systems. How can the phases of NLG systems be coordinated with the human writing process? And how should the parameters of the various phases be designed so that texts can be produced with good quality? (NLG systems in real life: writing on a meta level)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "4. As a result, we will formulate the requirements for the design of NLG systems that take into account the human writing process (Principles for designing NLG systems). These requirements ensure creating an environment in which the production of complex written texts is possible. The texts generated in this way should use the full potential of language and not just provide simple data descriptions 2 The different categories of NLG systems",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "In the basic reference work on NLG it is characterized as 'the subfield of artificial intelligence and computational linguistics that is concerned with the construction of computer systems that can produce understandable texts in English or other human languages from some underlying non-linguistic representation of information' (Reiter et al., 2000) . There are already a number of implemented applications for the data-to-text approach in different areas. They range from the media sector, where they have been a much-discussed topic as \"robot journalism\", to medical reports, business and finance reports or product descriptions in ecommerce. NLG systems are useful when large amounts of text are needed or information is only available in formats that are not easily understood (such as measurement data from medical examinations), and verbalisation facilitates or enables understanding.",
                "cite_spans": [
                    {
                        "start": 330,
                        "end": 351,
                        "text": "(Reiter et al., 2000)",
                        "ref_id": "BIBREF21"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "In this study, a further classification concerning the organization of NLG systems is to be discussed. On the one hand there are the so-called pipeline solutions that modularize the procedures and then execute the tasks (one after the other). The end-toend solutions on the other hand leave the modular approach behind and aim for end-to-end generation based on the successes of deep learning. They can be trained with (data, text) tuples that can be efficiently collected at scale (Castro Ferreira et al., 2019; Harkous et al., 2020) . Large pre-trained language models such as GPT-3 or BERT can be integrated into all of these solutions.",
                "cite_spans": [
                    {
                        "start": 482,
                        "end": 512,
                        "text": "(Castro Ferreira et al., 2019;",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 513,
                        "end": 534,
                        "text": "Harkous et al., 2020)",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "At present, end-to-end solutions are not yet suitable for commercial production of great amount of texts because they have two fundamental limitations: First, they are very domain-bound, so they can only generate texts for very limited segments. In addition, they lack semantic fidelity, this means how accurately the generated text conveys the meaning (Harkous et al., 2020) . As described, end-to-end systems based on deep learning combine all NLG steps in one function. This means that the only possible intervention is to select or edit the results (Gehrmann, 2020) . Due to this too tight restriction of interaction these approaches fall out of consideration for this research. Modular data-totext systems, on the other hand, offer more points of connection and reflect parallels between humans and systems in the text generation process.",
                "cite_spans": [
                    {
                        "start": 353,
                        "end": 375,
                        "text": "(Harkous et al., 2020)",
                        "ref_id": "BIBREF10"
                    },
                    {
                        "start": 553,
                        "end": 569,
                        "text": "(Gehrmann, 2020)",
                        "ref_id": "BIBREF8"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Since this study analyses the application under real-life conditions, the focus is on implementable solutions, not on academic NLG projects. In the commercial sector, rule-based pipeline solutions are established first and foremost, which differ in handling, architectures and purposes. Some of the solutions are offered as self-service, requiring limited or no programming skills. The leading companies in this fields are ARRIA NLG, Narrative Science, AX Semantics, Yseop and Automated Insights (Dale, 2020) .",
                "cite_spans": [
                    {
                        "start": 496,
                        "end": 508,
                        "text": "(Dale, 2020)",
                        "ref_id": "BIBREF4"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "There are different ways to structure the tasks and decisions of text generation. The most cited model for this is the NLG architecture constructed by Ehud Reiter and Robert Dale that performs tasks in sequence related to document planning, sentence planning and linguistic realization (Reiter et al., 2000) .",
                "cite_spans": [
                    {
                        "start": 286,
                        "end": 307,
                        "text": "(Reiter et al., 2000)",
                        "ref_id": "BIBREF21"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Overview of the NLG pipeline",
                "sec_num": "3"
            },
            {
                "text": "Content task Structure Task Document planning Content determination Document structuring",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Module",
                "sec_num": null
            },
            {
                "text": "Realisation Linguistic realisation Structure realisation Table 1 : Overview over the most important modules and tasks in the NLG pipeline (Reiter et al., 2000) The function of the Document Planner is to specify the text's content and structure based on domain and application knowledge about what information fits the specified communication goal and other generating objectives. In this module decisions are made about which information will be included (Content determination) and in what order this information will appear (Document structuring).",
                "cite_spans": [
                    {
                        "start": 138,
                        "end": 159,
                        "text": "(Reiter et al., 2000)",
                        "ref_id": "BIBREF21"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 57,
                        "end": 64,
                        "text": "Table 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Microplanning Lexicalisation Referring expression Generation Aggregation",
                "sec_num": null
            },
            {
                "text": "The task of the Microplanning component is to take the results of the Document Planning module and refine it to produce a more detailed text specifi-cation. At this point, sentences and paragraphs are planned (Aggregation) and the linguistic elements to be used to express the information are determined (Lexicalisation), i.e. which specific words or certain phrases are to be used. Within the Referring expression generation, it is decided which properties are used to describe an object unit, for example, a person's proper name and profession. It is therefore necessary to determine which properties are important so that the reader can identify the object.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Microplanning Lexicalisation Referring expression Generation Aggregation",
                "sec_num": null
            },
            {
                "text": "In the process of Surface Realisation, the system converts abstract representations of sentences into grammatically well-formed text (Linguistic realisation) and ensures that the abstract structures of sections and paragraphs are assembled as a document in the appropriate format.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Microplanning Lexicalisation Referring expression Generation Aggregation",
                "sec_num": null
            },
            {
                "text": "From the best-known model that illustrates how human writing functions at the cognitive level -the so called Flower-Hayes-Model (Flower and Hayes, 1981) -three important features can be derived that are characteristic of the human individual writing process:",
                "cite_spans": [
                    {
                        "start": 128,
                        "end": 152,
                        "text": "(Flower and Hayes, 1981)",
                        "ref_id": "BIBREF6"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "A model of the human writing process",
                "sec_num": "4"
            },
            {
                "text": "1. There are distinguishable cognitive processes.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "A model of the human writing process",
                "sec_num": "4"
            },
            {
                "text": "2. These processes are organized recursively.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "A model of the human writing process",
                "sec_num": "4"
            },
            {
                "text": "have an influence on further text production.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Text passages that have already been written",
                "sec_num": "3."
            },
            {
                "text": "The three distinguishable cognitive processes are controlled by a monitor. This central executive directs attention and switches from one sub-process to another.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Text passages that have already been written",
                "sec_num": "3."
            },
            {
                "text": "The first process is planning of a text, where information is collected and thoughts are made about the form and structure of the text. What should the text achieve? Whom does it address? What aspects, data, information should appear in it? It comprises three types of sub-operations: First there is generating, in which the writer retrieves information relevant to the writing task from longterm memory. Then there is organising, during which the most useful of the retrieved elements are arranged in a plan; finally the writer sets further goals to guide the writing (goal setting).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Text passages that have already been written",
                "sec_num": "3."
            },
            {
                "text": "After the planning follows the phase of linguistic implementation (translating). While many ideas in the planning phase are not really linguistically available, a kind of translation process now takes place during which the thoughts are translated into language: One now decides on the concrete vocabulary.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Text passages that have already been written",
                "sec_num": "3."
            },
            {
                "text": "The third main process is reviewing with its two sub-operations editing and revising. Now, the writer re-reads the text and aims to improve the quality of the written text by changing the text at the time it was written to correct errors, or fit the plans (editing). Or they intentionally revise the text to look for problems and errors at all levels of the text (revising).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Text passages that have already been written",
                "sec_num": "3."
            },
            {
                "text": "One of the most important findings of Flower and Hayes, which is also confirmed by later analyses of the writing processes, is the observation that writing is recursive: The writer jumps back and forth between the processes, again and again. There is no sequential proceeding in which one process is completed and then the next begins.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Recursiveness in writing",
                "sec_num": "4.1"
            },
            {
                "text": "In principle, it is possible to activate any process at any time, but it can be seen that the frequency and duration of the processes change throughout a writing session. The activation of translating remains constant while that of planning decreases and that of revision increases (Olive, 2004) . In the concrete act of writing, the recursive procedure shows itself in different facets:",
                "cite_spans": [
                    {
                        "start": 282,
                        "end": 295,
                        "text": "(Olive, 2004)",
                        "ref_id": "BIBREF17"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Recursiveness in writing",
                "sec_num": "4.1"
            },
            {
                "text": "\u2022 There is no fixed sequence of the individual operations. It seems that the individual writer develops certain patterns of sequences that remain relatively stable (Olive, 2004) .",
                "cite_spans": [
                    {
                        "start": 164,
                        "end": 177,
                        "text": "(Olive, 2004)",
                        "ref_id": "BIBREF17"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Recursiveness in writing",
                "sec_num": "4.1"
            },
            {
                "text": "\u2022 Individual activities always refer to each other and overlap.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Recursiveness in writing",
                "sec_num": "4.1"
            },
            {
                "text": "\u2022 All processes can be repeated as often as required.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Recursiveness in writing",
                "sec_num": "4.1"
            },
            {
                "text": "\u2022 Each formulation can be the trigger for a subsequent revision, which results in a new formulation, which in turn can be a trigger for another new formulation.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Recursiveness in writing",
                "sec_num": "4.1"
            },
            {
                "text": "Text passages written previously have influence on the further text and the arrangement of the processes. Reading and rereading the actual text is an important mental process in which the idea of the text is compared with the actual implementation. The deviations either lead to immediate changes in the written text or to a modification of the idea of the text -which, of course, in the further course of time influences both the text that is still being written and corrections of the pre-existing text passages.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Recursiveness in writing",
                "sec_num": "4.1"
            },
            {
                "text": "In general, technology and writing have always been interdependent: the writing tool and the writing medium influence writing in terms of how the problems at hand can be solved. In most writing settings today, the pen, pencil or typewriter has ceased to be the tool, and paper is no longer the medium. Rather, computers, tablets and smartphones with input functions and screens are the extended writing environment today (Mahlow and Dale, 2014) .",
                "cite_spans": [
                    {
                        "start": 421,
                        "end": 444,
                        "text": "(Mahlow and Dale, 2014)",
                        "ref_id": "BIBREF15"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "The writing environment in the narrower sense is the associated software. There have been and still are approaches to investigate which conditions serve the authors to write without interference and receive the appropriate support during the writing process.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "The investigation of the results of writing process research played an important role in this context (Sharples, 1999) . It was criticised that the writing tool and the medium were not included in former research. The most important results of the critique are, first, Sharples' (Sharples, 1999) re-evaluation of recursiveness and writing phases and second, the description of certain objects (external mental representations) as a bridge between the writer's ideas and the emerging text. He emphasises the biphasic nature of two activities within the writing process: engagement -this means the actual writing, where new material is created and reflection, the thinking (about the writing) where the generated material is revised. The two processes are separate and cannot occur simultaneously, forming cycles of engagement and reflection in writing (Sharples, 1999) .",
                "cite_spans": [
                    {
                        "start": 102,
                        "end": 118,
                        "text": "(Sharples, 1999)",
                        "ref_id": "BIBREF26"
                    },
                    {
                        "start": 279,
                        "end": 295,
                        "text": "(Sharples, 1999)",
                        "ref_id": "BIBREF26"
                    },
                    {
                        "start": 851,
                        "end": 867,
                        "text": "(Sharples, 1999)",
                        "ref_id": "BIBREF26"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "From these results, guidelines for the development of writing environment software were derived (Sharpies and Pemberton, 1990) with elaborating the following aspects:",
                "cite_spans": [
                    {
                        "start": 96,
                        "end": 126,
                        "text": "(Sharpies and Pemberton, 1990)",
                        "ref_id": "BIBREF25"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "\u2022 Because one cannot think about the structure of the text while writing, it is necessary to have a macrostructure (a kind of plan of the text), but this cannot be kept in our working memory. One needs an external representation of these macrostructures (Sharples, 1999 ).",
                "cite_spans": [
                    {
                        "start": 254,
                        "end": 269,
                        "text": "(Sharples, 1999",
                        "ref_id": "BIBREF26"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "\u2022 It must be possible to store mental representations of information (which can be in linear language or in other forms such as networks, mind maps, drawings or structures).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "\u2022 Writers need to be able to switch quickly between tasks (i.e. between notes, outline and linear text or spell check) this facilitates the interleaving of tasks.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "\u2022 Writers need to switch freely between different parts of the document as well, and should simultaneously be able to choose an appropriate level of focus. So, they should have an overview display and then be able to zoom in. At the same time, it has to be possible at all levels to delete or merge parts of the text or to change the order.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "Today there are a handful of software tools that take this non-linear writer-centred approach as their starting point (such as PageFour, Liquid Story Binder, RoughDraft (discontinued), Ulysses, Scrivener), but they tend to be used for specific professionalised, often narrative, writing (Bray, 2013) .",
                "cite_spans": [
                    {
                        "start": 287,
                        "end": 299,
                        "text": "(Bray, 2013)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "However, functions are built into conventional text processors as well that support individual phases of the writing process, such as the outline view, comment functions, text and grammar checks (Piotrowski and Mahlow, 2009) .",
                "cite_spans": [
                    {
                        "start": 195,
                        "end": 224,
                        "text": "(Piotrowski and Mahlow, 2009)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Requirements for writing software",
                "sec_num": "5"
            },
            {
                "text": "At this point, the phases of the human text production process and the modules of the NLG pipeline architecture are juxtaposed in order to find out which principles can be derived for an NLG system that is not designed for experts, but as a writing environment for the (automated) production of large numbers of texts.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG systems in real life: writing on a meta level",
                "sec_num": "6"
            },
            {
                "text": "The characteristic of this phase lies in the significance of alignment with the overall goals of the text: What are the interests of the target audience? What are the communication goals? This provides orientation for the selection of content and the structuring of the resulting text. The result depends on what goal is to be achieved with the texts and in which environment the text should be published. The editorial strategies as well as the narrative angles for the stories are developed.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: document planner -human writing: conceptual planning",
                "sec_num": "6.1"
            },
            {
                "text": "In individual writing, the writer derives such text assignments and keeps them either in long-term Figure 1 : The Flower-Hayes-Model, Flower and Hayes (1981) memory or in the form of a text brief or sample text. How detailed such specifications are worked out depends on the text assignment and the experience of the writer.",
                "cite_spans": [
                    {
                        "start": 134,
                        "end": 157,
                        "text": "Flower and Hayes (1981)",
                        "ref_id": "BIBREF6"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 99,
                        "end": 107,
                        "text": "Figure 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "NLG: document planner -human writing: conceptual planning",
                "sec_num": "6.1"
            },
            {
                "text": "In NLG systems the output of the document planner is a document plan which is a structured and ordered representation of messages. Often it is realized in form of a tree, whose leaf nodes are messages and whose internal nodes specify document elements such as paragraphs and sections and discourse relations between the elements. The representations of this plan are partly structural in nature, partly they are already connected with verbal elements (Reiter et al., 2000; Gatt and Krahmer, 2018) .",
                "cite_spans": [
                    {
                        "start": 451,
                        "end": 472,
                        "text": "(Reiter et al., 2000;",
                        "ref_id": "BIBREF21"
                    },
                    {
                        "start": 473,
                        "end": 496,
                        "text": "Gatt and Krahmer, 2018)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: document planner -human writing: conceptual planning",
                "sec_num": "6.1"
            },
            {
                "text": "Up to now humans were in most cases also responsible for designing handcrafted rules during the planning phase of automated text production, but there are some examples for developments of modelling genres with Machine Learning and statistics as long as there is a corpus of manual written text available for this specific case (Reiter and Williams, 2010) .",
                "cite_spans": [
                    {
                        "start": 328,
                        "end": 355,
                        "text": "(Reiter and Williams, 2010)",
                        "ref_id": "BIBREF23"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: document planner -human writing: conceptual planning",
                "sec_num": "6.1"
            },
            {
                "text": "At this point, it is worth considering how to transfer the author's implicit knowledge about the communication goal, text genre and document structure into explicit knowledge, such as rules, which can be applied to text generation. Many approaches are possible for the production of such a machineprocessable document plan by the writer: The necessary elements can be requested via a kind of questionnaire or forms can be filled out, based on briefing forms (Reiter and Williams, 2010) . Since in both areas a form of (internal) representation is created, that is still not translated into words, and for the reasons outlined above, namely that human Figure 2 ).",
                "cite_spans": [
                    {
                        "start": 458,
                        "end": 485,
                        "text": "(Reiter and Williams, 2010)",
                        "ref_id": "BIBREF23"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 651,
                        "end": 659,
                        "text": "Figure 2",
                        "ref_id": "FIGREF0"
                    }
                ],
                "eq_spans": [],
                "section": "NLG: document planner -human writing: conceptual planning",
                "sec_num": "6.1"
            },
            {
                "text": "Also in this phase, knowledge and information is inserted either by collecting data and doing research by the human author or by working with the database in NLG systems. In NLG systems data has to be filtered, mapped and combined to achieve the information needed. The results are semantic representations of information which are often expressed in logical or database languages (Gatt and Krahmer, 2018) . Commonly, in these systems the authors link particular data situations into abstract meaning which then can be used to trigger specific statements, phrases or document planning decisions. During production, data situations of the various data sets are then evaluated by the system and possible choices determined and executed upon. Especially compared to end-to-end neural systems, this makes sure that all aspects in the output are grounded in the underlying data.",
                "cite_spans": [
                    {
                        "start": 381,
                        "end": 405,
                        "text": "(Gatt and Krahmer, 2018)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: document planner -human writing: conceptual planning",
                "sec_num": "6.1"
            },
            {
                "text": "In this step, it is decided in which order information should appear in a text. As with planning what content is to be included, the orientation towards the reader group and the communication goal also applies here. In addition, there are some basic rhetorical rules and conventions for the individual text genres. For example, there is a rhetoric rule to place more general information at the top, while the details appear further back. In journalistic text forms on the other Figure 3 : This is a view of the logical structures of the statements and the first step to translating into language. (Data2Text Studio Interface, source: (Dou et al., 2018)) hand, the news, i.e. the special points, are mentioned first, while more general information comes later. There are some recent approaches to use machine learning techniques for content structuring, but since the text structure is very domain-oriented, its design is still produced on the basis of handwritten rules. This is where requirements for different levels of focus (Sharpies, 1992) come into play: It is advisable to be able to name or label the sentences and to represent them graphically so that they can be arranged by drag and drop, for example. Via the graphical representation, one can then access the assigned sentence and the appropriate data in order to be able to make changes at this level.",
                "cite_spans": [
                    {
                        "start": 634,
                        "end": 653,
                        "text": "(Dou et al., 2018))",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 1028,
                        "end": 1044,
                        "text": "(Sharpies, 1992)",
                        "ref_id": "BIBREF24"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 478,
                        "end": 486,
                        "text": "Figure 3",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "NLG: microplanning (aggregation)human writing: text structuring",
                "sec_num": "6.2"
            },
            {
                "text": "In this phase, the resulting nonverbal knowledge is translated into actual language. Now decisions have to be made about the words used and the syntax of the text.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "In NLG systems, one would basically have to transfer the non-linguistic concepts developed in the document planning phase directly into lexical elements. However, this is not easy for various reasons.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "First, the aspect of vagueness, which is tolerated in natural language, plays a major role here. Statements that are transferred as closely as possible directly from the data into words lead to a precision that is quickly perceived as unnatural in natural language. A certain degree of vagueness is necessary for expression in human languages.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "The second basic difficulty with this transferring task is that there are always several different ways to verbally describe a piece of information or an event. So there is not one solution for this task, but always multiple ones (Gatt and Krahmer, 2018) . For example, Reiter et al. (2005) discussed time expressions in the context of weather-forecast generation. A direct transfer of these time stamps into a record leads to the described overprecision (At 3:14 it was raining). Reiter et al. (2005) are also pointing out that e.g. a timestamp 00:00 could be expressed as late evening, midnight, or simply evening. Not surprisingly, humans show considerable variation in their lexical choices.",
                "cite_spans": [
                    {
                        "start": 230,
                        "end": 254,
                        "text": "(Gatt and Krahmer, 2018)",
                        "ref_id": "BIBREF7"
                    },
                    {
                        "start": 270,
                        "end": 290,
                        "text": "Reiter et al. (2005)",
                        "ref_id": "BIBREF22"
                    },
                    {
                        "start": 481,
                        "end": 501,
                        "text": "Reiter et al. (2005)",
                        "ref_id": "BIBREF22"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "Another consequence of this direct transmission would be the uniformity of expression, which is usually poorly tolerated in a text. If, for example, in weather texts a rise in temperature occurs several time and is expressed as follows:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "[time]+ [temp. rise in degrees] + {the temperature rose by}",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "The weather report for a day would look like this: In the morning the temperature rose by 4 degrees. In the afternoon the temperature rose by 5 degrees. In the the evening the temperature rose by -2 degrees.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "First the verbal expression rise for a negative rise would be fall. And in addition, such a formal structure would be identified very quickly and classified as unreadable. For this reason, several linguistic expressions must be available for a single pre-linguistic event, which are then selected by the system either randomly or based on a formulated condition derived, for example, from the communication goal or the rhetorical strategy. These linguistic variations also serve to ensure sufficient variance in the production of serial texts (see Figure  4) .",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 548,
                        "end": 558,
                        "text": "Figure  4)",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "The formulation of a larger set of different expressions for the data events is a task that in NLG systems still has to be performed by writers and is basically subject to the same principles as in the human language process.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "Unlike planning, the phase of translate is not related to spatial-visual functions of memory, but rather to phonological working memory: In principle, it is as if the writer now hears the words they write (Olive, 2004; Kellogg et al., 2007) . An abstract representation such as a plan or a formula does not provide support during this phase. For Figure 4 : This is a preview of multiple generated text for one data set to guarantee variance. (Data2Text Studio Interface, source: (Dou et al., 2018)) this reason, the user is always shown a real-time preview of what a possible instance of the statement would actually look like. Only in this form a statement can be heard.",
                "cite_spans": [
                    {
                        "start": 205,
                        "end": 218,
                        "text": "(Olive, 2004;",
                        "ref_id": "BIBREF17"
                    },
                    {
                        "start": 219,
                        "end": 240,
                        "text": "Kellogg et al., 2007)",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 479,
                        "end": 498,
                        "text": "(Dou et al., 2018))",
                        "ref_id": "BIBREF5"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 346,
                        "end": 354,
                        "text": "Figure 4",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "In this manner the user first develops an abstract formulaic representation of the text, then takes the intermediate step via preview and subsequently inserts the corrections into the formula (as an example of a separated preview table see Figures 3 and  4 ).",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 240,
                        "end": 256,
                        "text": "Figures 3 and  4",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "The sequence of this procedure, however, narrows the linguistic range of expression in comparison to the conventional formulation of an event. At this point, it is more suitable to give the writer the opportunity to phrase the sentence on the basis of a specific data set as if they were only producing an individual text. And only in a second step express the formula for this expression by providing the software with the labels and logics that it needs for further processing and that it cannot itself recognise on the basis of the text produced.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "At this stage, the application of an AI-based component is feasible. They can deliver suggestions based on e.g. keywords or paraphrases of the sentences created by the writer. Just as described earlier, the self-written text and the suggestions of the software take over the function of the already written text passages, which in turn can lead to new ideas for the next sentence or to revisions of previous text parts.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: micro planner -human writing: translating",
                "sec_num": "6.3"
            },
            {
                "text": "Linguistic Realisation is concerned with mapping the phrase specifications to the specific words and syntactic constructs which the target language provides such as making subject and verb agree, capitalizing the first letter of a sentence or building the correct plural of a noun. Most decisions in this stage are related to grammar (Reiter et al., 2000) . There are three approaches for implementing this task into NLG systems (Gatt and Krahmer, 2018) : Human-written templates that are easy to control, but require a lot of time and effort and offer only limited variability for texts; rule-based systems that make their choices on the basis of the grammar of the language; and statistic related solutions that rely on corpus data.",
                "cite_spans": [
                    {
                        "start": 334,
                        "end": 355,
                        "text": "(Reiter et al., 2000)",
                        "ref_id": "BIBREF21"
                    },
                    {
                        "start": 429,
                        "end": 453,
                        "text": "(Gatt and Krahmer, 2018)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: surface realizer -human writing: reviewing",
                "sec_num": "6.4"
            },
            {
                "text": "In the human writing process, an important part of these tasks is already accomplished in the verbalisation phase, but the validation of linguistic and grammatical accuracy takes place in the review phase. For checking syntax and grammar in the native language, the author usually relies on their linguistic intuition and looks up rules in case of doubt. In principle, however, they immediately recognise whether a concrete sentence is syntactically correct or not.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: surface realizer -human writing: reviewing",
                "sec_num": "6.4"
            },
            {
                "text": "It is less simple for them to assess correctness on the basis of abstract representations. For this reason, a separate review process for linguistic expression and correctness always has to be carried out on the basis of a sample of generated texts. In order to strategically adjust this review, it should be possible to compile this sample group on the basis of different criteria, such as the selection of specific evaluation data sets.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: surface realizer -human writing: reviewing",
                "sec_num": "6.4"
            },
            {
                "text": "It is noteworthy that NLG systems offer significant advantages in the review process over conventional word processors. Since they retain much more detailed linguistic information about the text, they can perform more targeted correcting operations than word processors. Thus, they fulfil the requirements that Piotrowski and Mahlow (2009) have formulated as to how a software must look like that supports the writer in their editing: (1) Specific views for highlighting linguistic phenomena, and (2) functions to perform operations on linguistic units.",
                "cite_spans": [
                    {
                        "start": 311,
                        "end": 339,
                        "text": "Piotrowski and Mahlow (2009)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: surface realizer -human writing: reviewing",
                "sec_num": "6.4"
            },
            {
                "text": "With NLG systems every change made in the text is automatically grammatically adjusted to ensure congruency: For example, changing the num-ber of the subject initiates changing the number of the finite verb and vice versa.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NLG: surface realizer -human writing: reviewing",
                "sec_num": "6.4"
            },
            {
                "text": "The following principles for the design of an NLG system can be derived from the observations presented above:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Principles for designing NLG systems",
                "sec_num": "7"
            },
            {
                "text": "1. Build modular systems in alignment with the writing processes: The modular design of conventional NLG systems suits the writer in that it can be used to provide them with the material and environment to support the specific stage of the writing process. Set up separate views for each main process, which are restricted to the processes in terms of their functionality.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Principles for designing NLG systems",
                "sec_num": "7"
            },
            {
                "text": "To comply with the recursiveness of human writing, it must be possible to edit each task at any time. On the one hand, this means that it must be possible to switch between tasks without any obstacles. And secondly, all changes within a task must be immediately passed on to all instances of the system.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Keep tasks flexible:",
                "sec_num": "2."
            },
            {
                "text": "3. Provide external (non-verbal) representations: In each phase, the writer must be able to draw on material that are not yet available as linear text. This includes not only overviews of the planning or outlines, but also the option of notes on the existing data material, formulated conditions, templates or text sections.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Keep tasks flexible:",
                "sec_num": "2."
            },
            {
                "text": "4. In the planning view give preference to visual information: This ranges from representations of the structure to illustrations of logics and data material.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Keep tasks flexible:",
                "sec_num": "2."
            },
            {
                "text": "5. Facilitate the possibilities for linguistic expression: The writer should always be able to write concrete sentences (without having to include formulas or other abstractions). Provide vocabulary or synonyms and ensure that the writer has the option of formulating multiple variations for the same statement.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Keep tasks flexible:",
                "sec_num": "2."
            },
            {
                "text": "6. Display instances of real text: The instance of a real text remains an important variable in the process. Only when real text is visible and editable linguistic creativity and grammatical correction can be adequately implemented. Even though this type of automated text production has different requirements as the production of an individually written text.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Keep tasks flexible:",
                "sec_num": "2."
            },
            {
                "text": "In rulebased data-to-text NLG systems, there is enough meta-information about the grammatical structure of the text that can be used for this task.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Enable linguistics-based editing:",
                "sec_num": "7."
            },
            {
                "text": "We showed that there are considerable similarities between the NLG modules and the writing phases of humans in terms of the tasks and decisions involved, which is a significant prerequisite for establishing these systems as a new extended writing technology. The analysis of these processes is of particular relevance in that quality assurance for data-to-text systems -whose goal is the mass generation of texts -is only attainable by optimizing the processes, since an evaluation of the entire output is not feasible.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "8"
            },
            {
                "text": "However, it also became clear that the human writing process has special features that need to be taken into account when designing NLG systems, especially the consistent and fast change between the processes and the distinctive cognitive activities that require access to different components of the human working memory (e.g. visio-spatial or phonological loop). To neglect these characteristics would mean confining the human involved in a linear process and to strict rules of formal language (i.e. code) to produce natural language texts. This kind of environment would impede the capacity of human writing and, with it, the quality of the text generated. In other words, it would stand in the way of a further successful development of the technology of writing which is to be expected in the course of adapting NLG systems in text production.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "8"
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "Writing with scrivener: A hopeful tale of disappearing tools, flatulence, and word processing redemption",
                "authors": [
                    {
                        "first": "Nancy",
                        "middle": [],
                        "last": "Bray",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "Computers and Composition",
                "volume": "30",
                "issue": "3",
                "pages": "197--210",
                "other_ids": {
                    "DOI": [
                        "10.1016/j.compcom.2013.07.002"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Nancy Bray. 2013. Writing with scrivener: A hopeful tale of disappearing tools, flatulence, and word pro- cessing redemption. Computers and Composition, 30(3):197-210.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "The robotic reporter",
                "authors": [
                    {
                        "first": "Matt",
                        "middle": [],
                        "last": "Carlson",
                        "suffix": ""
                    }
                ],
                "year": 2015,
                "venue": "Digital Journalism",
                "volume": "3",
                "issue": "3",
                "pages": "416--431",
                "other_ids": {
                    "DOI": [
                        "10.1080/21670811.2014.976412"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Matt Carlson. 2015. The robotic reporter. Digital Jour- nalism, 3(3):416-431.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Neural datato-text generation: A comparison between pipeline and end-to",
                "authors": [
                    {
                        "first": "Chris",
                        "middle": [],
                        "last": "Thiago Castro Ferreira",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Van Der Lee",
                        "suffix": ""
                    },
                    {
                        "first": "Emiel",
                        "middle": [],
                        "last": "Emiel Van Miltenburg",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Krahmer",
                        "suffix": ""
                    }
                ],
                "year": 2019,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Thiago Castro Ferreira, Chris van der Lee, Emiel van Miltenburg, and Emiel Krahmer. 2019. Neural data- to-text generation: A comparison between pipeline and end-to-end architectures.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Creative writing with a machine in the loop: Case studies on slogans and stories",
                "authors": [
                    {
                        "first": "Elizabeth",
                        "middle": [],
                        "last": "Clark",
                        "suffix": ""
                    },
                    {
                        "first": "Anne",
                        "middle": [
                            "Spencer"
                        ],
                        "last": "Ross",
                        "suffix": ""
                    },
                    {
                        "first": "Chenhao",
                        "middle": [],
                        "last": "Tan",
                        "suffix": ""
                    },
                    {
                        "first": "Yangfeng",
                        "middle": [],
                        "last": "Ji",
                        "suffix": ""
                    },
                    {
                        "first": "Noah",
                        "middle": [
                            "A"
                        ],
                        "last": "Smith",
                        "suffix": ""
                    }
                ],
                "year": 2018,
                "venue": "23rd International Conference on Intelligent User Interfaces, IUI '18",
                "volume": "",
                "issue": "",
                "pages": "329--340",
                "other_ids": {
                    "DOI": [
                        "10.1145/3172944.3172983"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Elizabeth Clark, Anne Spencer Ross, Chenhao Tan, Yangfeng Ji, and Noah A. Smith. 2018. Creative writing with a machine in the loop: Case studies on slogans and stories. In 23rd International Confer- ence on Intelligent User Interfaces, IUI '18, page 329-340, New York, NY, USA. Association for Com- puting Machinery.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "Natural language generation: The commercial state of the art in 2020",
                "authors": [
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Dale",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "Natural Language Engineering",
                "volume": "26",
                "issue": "4",
                "pages": "481--487",
                "other_ids": {
                    "DOI": [
                        "10.1017/S135132492000025X"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Robert Dale. 2020. Natural language generation: The commercial state of the art in 2020. Natural Lan- guage Engineering, 26(4):481-487.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "Data2Text studio: Automated text generation from structured data",
                "authors": [
                    {
                        "first": "Longxu",
                        "middle": [],
                        "last": "Dou",
                        "suffix": ""
                    },
                    {
                        "first": "Guanghui",
                        "middle": [],
                        "last": "Qin",
                        "suffix": ""
                    },
                    {
                        "first": "Jinpeng",
                        "middle": [],
                        "last": "Wang",
                        "suffix": ""
                    },
                    {
                        "first": "Jin-Ge",
                        "middle": [],
                        "last": "Yao",
                        "suffix": ""
                    },
                    {
                        "first": "Chin-Yew",
                        "middle": [],
                        "last": "Lin",
                        "suffix": ""
                    }
                ],
                "year": 2018,
                "venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations",
                "volume": "",
                "issue": "",
                "pages": "13--18",
                "other_ids": {
                    "DOI": [
                        "10.18653/v1/D18-2003"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Longxu Dou, Guanghui Qin, Jinpeng Wang, Jin-Ge Yao, and Chin-Yew Lin. 2018. Data2Text studio: Automated text generation from structured data. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 13-18, Brussels, Belgium. Association for Computational Linguistics.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "A Cognitive Process Theory of Writing. College Composition and Communication",
                "authors": [
                    {
                        "first": "Linda",
                        "middle": [],
                        "last": "Flower",
                        "suffix": ""
                    },
                    {
                        "first": "John",
                        "middle": [
                            "R"
                        ],
                        "last": "Hayes",
                        "suffix": ""
                    }
                ],
                "year": 1981,
                "venue": "",
                "volume": "32",
                "issue": "",
                "pages": "365--387",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Linda Flower and John R. Hayes. 1981. A Cognitive Process Theory of Writing. College Composition and Communication, 32(4):365-387.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Survey of the state of the art in natural language generation: Core tasks, applications and evaluation",
                "authors": [
                    {
                        "first": "Albert",
                        "middle": [],
                        "last": "Gatt",
                        "suffix": ""
                    },
                    {
                        "first": "Emiel",
                        "middle": [],
                        "last": "Krahmer",
                        "suffix": ""
                    }
                ],
                "year": 2018,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Albert Gatt and Emiel Krahmer. 2018. Survey of the state of the art in natural language generation: Core tasks, applications and evaluation.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Human-AI collaboration for natural language generation with interpretable neural networks",
                "authors": [
                    {
                        "first": "",
                        "middle": [],
                        "last": "Sebastian Gehrmann",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Sebastian Gehrmann. 2020. Human-AI collaboration for natural language generation with interpretable neural networks. Ph.D. thesis, Harvard University.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "Repairing the cracked foundation: A survey of obstacles in evaluation practices for generated text",
                "authors": [
                    {
                        "first": "Sebastian",
                        "middle": [],
                        "last": "Gehrmann",
                        "suffix": ""
                    },
                    {
                        "first": "Elizabeth",
                        "middle": [],
                        "last": "Clark",
                        "suffix": ""
                    },
                    {
                        "first": "Thibault",
                        "middle": [],
                        "last": "Sellam",
                        "suffix": ""
                    }
                ],
                "year": 2022,
                "venue": "ArXiv",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Sebastian Gehrmann, Elizabeth Clark, and Thibault Sel- lam. 2022. Repairing the cracked foundation: A sur- vey of obstacles in evaluation practices for generated text. ArXiv, abs/2202.06935.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Have your text and use it too! end-to-end neural datato-text generation with semantic fidelity",
                "authors": [
                    {
                        "first": "Hamza",
                        "middle": [],
                        "last": "Harkous",
                        "suffix": ""
                    },
                    {
                        "first": "Isabel",
                        "middle": [],
                        "last": "Groves",
                        "suffix": ""
                    },
                    {
                        "first": "Amir",
                        "middle": [],
                        "last": "Saffari",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "Proceedings of the 28th International Conference on Computational Linguistics",
                "volume": "",
                "issue": "",
                "pages": "2410--2424",
                "other_ids": {
                    "DOI": [
                        "10.18653/v1/2020.coling-main.218"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Hamza Harkous, Isabel Groves, and Amir Saffari. 2020. Have your text and use it too! end-to-end neural data- to-text generation with semantic fidelity. In Proceed- ings of the 28th International Conference on Com- putational Linguistics, pages 2410-2424, Barcelona, Spain (Online). International Committee on Compu- tational Linguistics.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "Tool vs. agent: attributing agency to natural language generation systems",
                "authors": [
                    {
                        "first": "Leah",
                        "middle": [],
                        "last": "Henrickson",
                        "suffix": ""
                    }
                ],
                "year": 2018,
                "venue": "Digital Creativity",
                "volume": "29",
                "issue": "2-3",
                "pages": "182--190",
                "other_ids": {
                    "DOI": [
                        "10.1080/14626268.2018.1482924"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Leah Henrickson. 2018. Tool vs. agent: attributing agency to natural language generation systems. Digi- tal Creativity, 29(2-3):182-190.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Emiel van Miltenburg, Sashank Santhanam, and Verena Rieser. 2020. Twenty years of confusion in human evaluation: Nlg needs evaluation sheets and standardised definitions",
                "authors": [
                    {
                        "first": "David",
                        "middle": [
                            "M"
                        ],
                        "last": "Howcroft",
                        "suffix": ""
                    },
                    {
                        "first": "Anya",
                        "middle": [],
                        "last": "Belz",
                        "suffix": ""
                    },
                    {
                        "first": "Miruna",
                        "middle": [],
                        "last": "Clinciu",
                        "suffix": ""
                    },
                    {
                        "first": "Dimitra",
                        "middle": [],
                        "last": "Gkatzia",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Sadid",
                        "suffix": ""
                    },
                    {
                        "first": "Saad",
                        "middle": [],
                        "last": "Hasan",
                        "suffix": ""
                    },
                    {
                        "first": "Simon",
                        "middle": [],
                        "last": "Mahamood",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Mille",
                        "suffix": ""
                    }
                ],
                "year": null,
                "venue": "INLG",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "David M. Howcroft, Anya Belz, Miruna Clinciu, Dimi- tra Gkatzia, Sadid A. Hasan, Saad Mahamood, Simon Mille, Emiel van Miltenburg, Sashank Santhanam, and Verena Rieser. 2020. Twenty years of confusion in human evaluation: Nlg needs evaluation sheets and standardised definitions. In INLG.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "Verbal, visual, and spatial working memory in written language production",
                "authors": [
                    {
                        "first": "Ronald",
                        "middle": [
                            "T"
                        ],
                        "last": "Kellogg",
                        "suffix": ""
                    },
                    {
                        "first": "Thierry",
                        "middle": [],
                        "last": "Olive",
                        "suffix": ""
                    },
                    {
                        "first": "Annie",
                        "middle": [],
                        "last": "Piolat",
                        "suffix": ""
                    }
                ],
                "year": 2007,
                "venue": "Acta Psychologica",
                "volume": "124",
                "issue": "3",
                "pages": "382--397",
                "other_ids": {
                    "DOI": [
                        "10.1016/j.actpsy.2006.02.005"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Ronald T. Kellogg, Thierry Olive, and Annie Piolat. 2007. Verbal, visual, and spatial working memory in written language production. Acta Psychologica, 124(3):382-397.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Automation, journalism, and human-machine communication: Rethinking roles and relationships of humans and machines in news",
                "authors": [
                    {
                        "first": "Seth",
                        "middle": [],
                        "last": "Lewis",
                        "suffix": ""
                    },
                    {
                        "first": "Andrea",
                        "middle": [],
                        "last": "Guzman",
                        "suffix": ""
                    },
                    {
                        "first": "Thomas",
                        "middle": [],
                        "last": "Schmidt",
                        "suffix": ""
                    }
                ],
                "year": 2019,
                "venue": "Digital Journalism",
                "volume": "7",
                "issue": "",
                "pages": "1--19",
                "other_ids": {
                    "DOI": [
                        "10.1080/21670811.2019.1577147"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Seth Lewis, Andrea Guzman, and Thomas Schmidt. 2019. Automation, journalism, and human-machine communication: Rethinking roles and relationships of humans and machines in news. Digital Journalism, 7:1-19.",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "Production Media: Writing as Using Tools in Media Convergent Environments",
                "authors": [
                    {
                        "first": "Cerstin",
                        "middle": [],
                        "last": "Mahlow",
                        "suffix": ""
                    },
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Dale",
                        "suffix": ""
                    }
                ],
                "year": 2014,
                "venue": "Handbook of Writing and Text Production",
                "volume": "10",
                "issue": "",
                "pages": "209--230",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Cerstin Mahlow and Robert Dale. 2014. Production Media: Writing as Using Tools in Media Conver- gent Environments. In Eva-Maria Jakobs and Daniel Perrin, editors, Handbook of Writing and Text Pro- duction, volume 10 of Handbooks of Applied Lin- guistics, pages 209-230. De Gruyter Mouton, Berlin, Germany.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "Synthetic literature: Writing science fiction in a co-creative process",
                "authors": [
                    {
                        "first": "Enrique",
                        "middle": [],
                        "last": "Manjavacas",
                        "suffix": ""
                    },
                    {
                        "first": "Folgert",
                        "middle": [],
                        "last": "Karsdorp",
                        "suffix": ""
                    },
                    {
                        "first": "Ben",
                        "middle": [],
                        "last": "Burtenshaw",
                        "suffix": ""
                    },
                    {
                        "first": "Mike",
                        "middle": [],
                        "last": "Kestemont",
                        "suffix": ""
                    }
                ],
                "year": 2017,
                "venue": "Proceedings of the Workshop on Computational Creativity in Natural Language Generation (CC-NLG 2017)",
                "volume": "",
                "issue": "",
                "pages": "29--37",
                "other_ids": {
                    "DOI": [
                        "10.18653/v1/W17-3904"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Enrique Manjavacas, Folgert Karsdorp, Ben Burten- shaw, and Mike Kestemont. 2017. Synthetic liter- ature: Writing science fiction in a co-creative pro- cess. In Proceedings of the Workshop on Compu- tational Creativity in Natural Language Generation (CC-NLG 2017), pages 29-37, Santiago de Com- postela, Spain. Association for Computational Lin- guistics.",
                "links": null
            },
            "BIBREF17": {
                "ref_id": "b17",
                "title": "Working memory in writing: Empirical evidence from the dual-task technique",
                "authors": [
                    {
                        "first": "Thierry",
                        "middle": [],
                        "last": "Olive",
                        "suffix": ""
                    }
                ],
                "year": 2004,
                "venue": "European Psychologist",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {
                    "DOI": [
                        "10.1027/1016-9040.9.1.32"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Thierry Olive. 2004. Working memory in writing: Em- pirical evidence from the dual-task technique. Euro- pean Psychologist.",
                "links": null
            },
            "BIBREF18": {
                "ref_id": "b18",
                "title": "Schreiben: von intuitiven zu professionellen Schreibstrategien",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Perrin",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D. Perrin. 2002. Schreiben: von intuitiven zu profes- sionellen Schreibstrategien. Westdt. Verlag.",
                "links": null
            },
            "BIBREF19": {
                "ref_id": "b19",
                "title": "Wie Journalisten schreiben. Ergebnisse angewandter Schreibprozessforschung",
                "authors": [
                    {
                        "first": "Daniel",
                        "middle": [],
                        "last": "Perrin",
                        "suffix": ""
                    }
                ],
                "year": 2001,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Daniel Perrin. 2001. Wie Journalisten schreiben. Ergeb- nisse angewandter Schreibprozessforschung. UVK Verlag.",
                "links": null
            },
            "BIBREF20": {
                "ref_id": "b20",
                "title": "Linguistic editing support",
                "authors": [
                    {
                        "first": "Michael",
                        "middle": [],
                        "last": "Piotrowski",
                        "suffix": ""
                    },
                    {
                        "first": "Cerstin",
                        "middle": [],
                        "last": "Mahlow",
                        "suffix": ""
                    }
                ],
                "year": 2009,
                "venue": "Proceedings of the 9th ACM Symposium on Document Engineering",
                "volume": "",
                "issue": "",
                "pages": "214--217",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Michael Piotrowski and Cerstin Mahlow. 2009. Lin- guistic editing support. In Proceedings of the 9th ACM Symposium on Document Engineering, pages 214-217.",
                "links": null
            },
            "BIBREF21": {
                "ref_id": "b21",
                "title": "Building natural language generation systems",
                "authors": [
                    {
                        "first": "Ehud",
                        "middle": [],
                        "last": "Reiter",
                        "suffix": ""
                    },
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Dale",
                        "suffix": ""
                    },
                    {
                        "first": "Zhiwei",
                        "middle": [],
                        "last": "Feng",
                        "suffix": ""
                    }
                ],
                "year": 2000,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Ehud Reiter, Robert Dale, and Zhiwei Feng. 2000. Building natural language generation systems, vol- ume 33. MIT Press.",
                "links": null
            },
            "BIBREF22": {
                "ref_id": "b22",
                "title": "Choosing words in computergenerated weather forecasts",
                "authors": [
                    {
                        "first": "Ehud",
                        "middle": [],
                        "last": "Reiter",
                        "suffix": ""
                    },
                    {
                        "first": "Somayajulu",
                        "middle": [],
                        "last": "Sripada",
                        "suffix": ""
                    },
                    {
                        "first": "Jim",
                        "middle": [],
                        "last": "Hunter",
                        "suffix": ""
                    },
                    {
                        "first": "Jin",
                        "middle": [],
                        "last": "Yu",
                        "suffix": ""
                    },
                    {
                        "first": "Ian",
                        "middle": [],
                        "last": "Davy",
                        "suffix": ""
                    }
                ],
                "year": 2005,
                "venue": "Artificial Intelligence",
                "volume": "167",
                "issue": "",
                "pages": "137--169",
                "other_ids": {
                    "DOI": [
                        "10.1016/j.artint.2005.06.006"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Ehud Reiter, Somayajulu Sripada, Jim Hunter, Jin Yu, and Ian Davy. 2005. Choosing words in computer- generated weather forecasts. Artificial Intelligence, 167:137-169.",
                "links": null
            },
            "BIBREF23": {
                "ref_id": "b23",
                "title": "Generating texts in different styles",
                "authors": [
                    {
                        "first": "Ehud",
                        "middle": [],
                        "last": "Reiter",
                        "suffix": ""
                    },
                    {
                        "first": "Sandra",
                        "middle": [],
                        "last": "Williams",
                        "suffix": ""
                    }
                ],
                "year": 2010,
                "venue": "The Structure of Style",
                "volume": "",
                "issue": "",
                "pages": "59--75",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Ehud Reiter and Sandra Williams. 2010. Generating texts in different styles. In The Structure of Style, pages 59-75. Springer.",
                "links": null
            },
            "BIBREF24": {
                "ref_id": "b24",
                "title": "Representing writing: External representations. Computers and Writing: State of the Art",
                "authors": [
                    {
                        "first": "Mike",
                        "middle": [],
                        "last": "Sharpies",
                        "suffix": ""
                    }
                ],
                "year": 1992,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Mike Sharpies. 1992. Representing writing: External representations. Computers and Writing: State of the Art.",
                "links": null
            },
            "BIBREF25": {
                "ref_id": "b25",
                "title": "Starting from the writer: Guidelines for the design of user-centred document processors",
                "authors": [
                    {
                        "first": "Mike",
                        "middle": [],
                        "last": "Sharpies",
                        "suffix": ""
                    },
                    {
                        "first": "Lyn",
                        "middle": [],
                        "last": "Pemberton",
                        "suffix": ""
                    }
                ],
                "year": 1990,
                "venue": "Computer Assisted Language Learning",
                "volume": "2",
                "issue": "1",
                "pages": "37--57",
                "other_ids": {
                    "DOI": [
                        "10.1080/0958822900020104"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Mike Sharpies and Lyn Pemberton. 1990. Starting from the writer: Guidelines for the design of user-centred document processors. Computer Assisted Language Learning, 2(1):37-57.",
                "links": null
            },
            "BIBREF26": {
                "ref_id": "b26",
                "title": "How we write. Writing as creative design",
                "authors": [
                    {
                        "first": "Mike",
                        "middle": [],
                        "last": "Sharples",
                        "suffix": ""
                    }
                ],
                "year": 1999,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Mike Sharples. 1999. How we write. Writing as creative design. Routledge.",
                "links": null
            },
            "BIBREF27": {
                "ref_id": "b27",
                "title": "Qualit\u00e4tsmanagement in Redaktionen, number 1 in Aktuell : Studien zum Journalismus",
                "authors": [
                    {
                        "first": "Vinzenz",
                        "middle": [],
                        "last": "Wyss",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "89--105",
                "other_ids": {
                    "DOI": [
                        "10.5771/9783845236933-89"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Vinzenz Wyss. 2013. Qualit\u00e4tsmanagement in Redak- tionen, number 1 in Aktuell : Studien zum Journalis- mus, pages 89-105. Neuberger, Christoph and Meier, Klaus, Baden-Baden.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "uris": null,
                "num": null,
                "type_str": "figure",
                "text": "This is the main view in the writing software Scrivener. This is an example how a graphic representation with verbal elements can look. On the right there is the option to label and annotate the text parts. (Source: https://www.literatureandlatte. com/scrivener/overview) mental representations of the document structure are often visual, graphic solutions are a suitable choice. A good example for this is the main page of the writing software Scrivener (Literature and Latte) (see"
            }
        }
    }
}